[
{
    "branch": "main",
    "index": "0",
    "phase": "pre",
    "playbook": "github.com/osism/zuul-config/playbooks/base/pre.yaml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-03-21T00:02:00.773553Z",
                    "start": "2026-03-21T00:00:18.562375Z"
                },
                "id": "0242ac1f-6606-1af1-8b27-000000000002",
                "name": "Base pre"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "include_role",
                            "changed": false,
                            "include_args": {
                                "name": "set-zuul-log-path-fact"
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-00000000000a",
                        "name": "emit-job-header",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/emit-job-header"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:18.671283Z",
                            "start": "2026-03-21T00:00:18.634892Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-00000000000c",
                        "name": "Setup log path fact"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "zuul_log_path": "a32/osism/a3240bd8b4c74e6f8c9446925cc2fed6"
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-000000000087",
                        "name": "set-zuul-log-path-fact",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/set-zuul-log-path-fact"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:18.815150Z",
                            "start": "2026-03-21T00:00:18.716629Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-000000000089",
                        "name": "Set log path for a build"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "debug",
                            "changed": false,
                            "msg": "# Job Information\nAnsible Version: 2.16.14\nJob: testbed-deploy-current-in-a-nutshell-with-tempest-ubuntu-24.04\nPipeline: periodic-midnight\nExecutor: 521e9411259a\nTriggered by: https://github.com/osism/testbed\nEvent ID: 7b65fefe337148519e9759ea87942a86\n"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-00000000000a",
                        "name": "emit-job-header",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/emit-job-header"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:18.896625Z",
                            "start": "2026-03-21T00:00:18.837201Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-00000000000d",
                        "name": "Print job information"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "debug",
                            "changed": false,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_item",
                                    "changed": false,
                                    "failed": false,
                                    "msg": "# Node Information\nInventory Hostname: orchestrator\nHostname: zuul-static-regiocloud-infra-1\nUsername: zuul-testbed04\nDistro: Debian 12.13\nProvider: static-testbed\nRegion: \nLabel: testbed-orchestrator\nProduct Name: OpenStack Nova\nInterface IP: 81.163.193.140\n",
                                    "zj_item": "orchestrator"
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-00000000000a",
                        "name": "emit-job-header",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/emit-job-header"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:19.136909Z",
                            "start": "2026-03-21T00:00:18.902659Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-00000000000e",
                        "name": "Print node information"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": true,
                            "diff": {
                                "after": {
                                    "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/logs/zuul-info",
                                    "state": "directory"
                                },
                                "before": {
                                    "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/logs/zuul-info",
                                    "state": "absent"
                                }
                            },
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": 493,
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/logs/zuul-info",
                                    "recurse": false,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0755",
                            "owner": "root",
                            "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/logs/zuul-info",
                            "size": 4096,
                            "state": "directory",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-000000000011",
                        "name": "log-inventory",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/log-inventory"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:20.195911Z",
                            "start": "2026-03-21T00:00:19.146967Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-000000000013",
                        "name": "Ensure Zuul Ansible directory exists"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "copy",
                            "changed": true,
                            "checksum": "77470e381c3118747bd1c9d56bb12e58b4f2b14a",
                            "dest": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/logs/zuul-info/inventory.yaml",
                            "diff": [],
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_original_basename": "inventory.yaml",
                                    "attributes": null,
                                    "backup": false,
                                    "checksum": "77470e381c3118747bd1c9d56bb12e58b4f2b14a",
                                    "content": null,
                                    "dest": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/logs/zuul-info",
                                    "directory_mode": null,
                                    "follow": false,
                                    "force": true,
                                    "group": null,
                                    "local_follow": null,
                                    "mode": 420,
                                    "owner": null,
                                    "remote_src": null,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/.ansible/tmp/ansible-tmp-1774051220.273074-34-226251784255386/source",
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "md5sum": "694fd5fb506ba31f222c00c0aff8c634",
                            "mode": "0644",
                            "owner": "root",
                            "size": 8206,
                            "src": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/.ansible/tmp/ansible-tmp-1774051220.273074-34-226251784255386/source",
                            "state": "file",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-000000000011",
                        "name": "log-inventory",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/log-inventory"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:22.361364Z",
                            "start": "2026-03-21T00:00:20.202307Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-000000000014",
                        "name": "Copy ansible inventory to logs dir"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "stat",
                            "changed": false,
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "checksum_algorithm": "sha1",
                                    "follow": false,
                                    "get_attributes": true,
                                    "get_checksum": true,
                                    "get_mime": true,
                                    "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/a3240bd8b4c74e6f8c9446925cc2fed6_id_rsa"
                                }
                            },
                            "stat": {
                                "exists": false
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:22.826235Z",
                            "start": "2026-03-21T00:00:22.374481Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-000000000018",
                        "name": "Check to see if ssh key was already created for this build"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "include_tasks",
                            "changed": false,
                            "include": "create-key-and-replace.yaml",
                            "include_args": {}
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:22.889735Z",
                            "start": "2026-03-21T00:00:22.831725Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-000000000019",
                        "name": "Create a new key in workspace based on build UUID"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": true,
                            "cmd": [
                                "ssh-keygen",
                                "-t",
                                "rsa",
                                "-N",
                                "",
                                "-C",
                                "zuul-build-sshkey",
                                "-f",
                                "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/a3240bd8b4c74e6f8c9446925cc2fed6_id_rsa",
                                "-b",
                                "3072"
                            ],
                            "delta": "0:00:01.477876",
                            "end": "2026-03-21 00:00:25.365428",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh-keygen -t rsa -N '' -C 'zuul-build-sshkey' -f /var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/a3240bd8b4c74e6f8c9446925cc2fed6_id_rsa -b 3072",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-1af1-8b27-0000000000aa-0-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-21 00:00:23.887552",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "Generating public/private rsa key pair.\nYour identification has been saved in /var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/a3240bd8b4c74e6f8c9446925cc2fed6_id_rsa\nYour public key has been saved in /var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/a3240bd8b4c74e6f8c9446925cc2fed6_id_rsa.pub\nThe key fingerprint is:\nSHA256:TkDzE75PHppPE5gNRppHB0lhiLIaaJgRl++HrqHCLnQ zuul-build-sshkey\nThe key's randomart image is:\n+---[RSA 3072]----+\n|.... .ooO+.      |\n|..o ...X.o       |\n|oo +  + B        |\n|=.. .  + B       |\n|.o . .  S =      |\n|.. Eo .o * o     |\n|o o. .  + =      |\n|oo ..    o .     |\n|=...      .      |\n+----[SHA256]-----+",
                            "stdout_lines": [
                                "Generating public/private rsa key pair.",
                                "Your identification has been saved in /var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/a3240bd8b4c74e6f8c9446925cc2fed6_id_rsa",
                                "Your public key has been saved in /var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/a3240bd8b4c74e6f8c9446925cc2fed6_id_rsa.pub",
                                "The key fingerprint is:",
                                "SHA256:TkDzE75PHppPE5gNRppHB0lhiLIaaJgRl++HrqHCLnQ zuul-build-sshkey",
                                "The key's randomart image is:",
                                "+---[RSA 3072]----+",
                                "|.... .ooO+.      |",
                                "|..o ...X.o       |",
                                "|oo +  + B        |",
                                "|=.. .  + B       |",
                                "|.o . .  S =      |",
                                "|.. Eo .o * o     |",
                                "|o o. .  + =      |",
                                "|oo ..    o .     |",
                                "|=...      .      |",
                                "+----[SHA256]-----+"
                            ],
                            "zuul_log_id": "0242ac1f-6606-1af1-8b27-0000000000aa-0-orchestrator"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:25.454014Z",
                            "start": "2026-03-21T00:00:22.945677Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-0000000000aa",
                        "name": "Create Temp SSH key"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "include_tasks",
                            "changed": false,
                            "include": "remote-linux.yaml",
                            "include_args": {}
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:25.509809Z",
                            "start": "2026-03-21T00:00:25.459545Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-0000000000ab",
                        "name": "Remote setup ssh keys (linux)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "lineinfile",
                            "changed": false,
                            "false_condition": "zuul_build_sshkey_cleanup",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:25.616196Z",
                            "start": "2026-03-21T00:00:25.550020Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-0000000000ca",
                        "name": "Remove previously added zuul-build-sshkey"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "authorized_key",
                            "changed": true,
                            "comment": null,
                            "exclusive": false,
                            "follow": false,
                            "invocation": {
                                "module_args": {
                                    "changed": true,
                                    "comment": null,
                                    "exclusive": false,
                                    "follow": false,
                                    "key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDiFU4X6bBG4GOayutbGDh+9mZMjWvhGFfP2Ynam+67TPObfnNalmbdJk06dNX/TYaC+yECnjjU2QtjNemgB1Of/27WgEMDtf2N+7DmrkWuvstUf0h1Qz2h5b0CxoqRCje/cCp4yqEFtqUuGRjZoGYEFZBIN1kvfPDNLFPyPYPfcxEgMtuwS87HTBdyWTHJu2NYPKyVjSLKpkiwtL19MyrYhjkD+Xlvaf6BJFCpdIbxa3muVVRkFr4vLWC+Rhu79Fk4XCy6MvyhU1VkAiv7qvjfE8KyeVnDvLy5nWYfHtyIa5rPDtwkoeHJdfsf4UN0qm2oNOHlMcD1fmg3jDUdDjpXYy8hLhiC0qKMRIiRTokNU+zw1rTDJSr78tAzap9ouJ6D0Tq10mqBZGUEpyM7n37IFOPU2fQmeoSvGQiWIl0dporJnK+QmiDEbs742C3DFQaEPXaJ5K4Gm6LcxeGgsqCzFXjGPOtIG98tST2dNvyzJX+xoi/i7/ALp03+FFZccaM= zuul-build-sshkey",
                                    "key_options": null,
                                    "keyfile": "/home/zuul-testbed04/.ssh/authorized_keys",
                                    "manage_dir": true,
                                    "path": null,
                                    "state": "present",
                                    "user": "zuul-testbed04",
                                    "validate_certs": true
                                }
                            },
                            "key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDiFU4X6bBG4GOayutbGDh+9mZMjWvhGFfP2Ynam+67TPObfnNalmbdJk06dNX/TYaC+yECnjjU2QtjNemgB1Of/27WgEMDtf2N+7DmrkWuvstUf0h1Qz2h5b0CxoqRCje/cCp4yqEFtqUuGRjZoGYEFZBIN1kvfPDNLFPyPYPfcxEgMtuwS87HTBdyWTHJu2NYPKyVjSLKpkiwtL19MyrYhjkD+Xlvaf6BJFCpdIbxa3muVVRkFr4vLWC+Rhu79Fk4XCy6MvyhU1VkAiv7qvjfE8KyeVnDvLy5nWYfHtyIa5rPDtwkoeHJdfsf4UN0qm2oNOHlMcD1fmg3jDUdDjpXYy8hLhiC0qKMRIiRTokNU+zw1rTDJSr78tAzap9ouJ6D0Tq10mqBZGUEpyM7n37IFOPU2fQmeoSvGQiWIl0dporJnK+QmiDEbs742C3DFQaEPXaJ5K4Gm6LcxeGgsqCzFXjGPOtIG98tST2dNvyzJX+xoi/i7/ALp03+FFZccaM= zuul-build-sshkey",
                            "key_options": null,
                            "keyfile": "/home/zuul-testbed04/.ssh/authorized_keys",
                            "manage_dir": true,
                            "path": null,
                            "state": "present",
                            "user": "zuul-testbed04",
                            "validate_certs": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:26.785417Z",
                            "start": "2026-03-21T00:00:25.625060Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-0000000000cb",
                        "name": "Enable access via build key on all nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": false,
                            "diff": {
                                "after": {
                                    "path": "/home/zuul-testbed04/.ssh"
                                },
                                "before": {
                                    "path": "/home/zuul-testbed04/.ssh"
                                }
                            },
                            "gid": 1005,
                            "group": "zuul-testbed04",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": 448,
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/home/zuul-testbed04/.ssh",
                                    "recurse": false,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0700",
                            "owner": "zuul-testbed04",
                            "path": "/home/zuul-testbed04/.ssh",
                            "size": 4096,
                            "state": "directory",
                            "uid": 1004
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:27.129150Z",
                            "start": "2026-03-21T00:00:26.790667Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-0000000000cc",
                        "name": "Make sure user has a .ssh"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "copy",
                            "changed": false,
                            "dest": "/home/zuul-testbed04/.ssh/id_rsa",
                            "invocation": {
                                "dest": "~/.ssh/id_rsa",
                                "force": false,
                                "mode": 384,
                                "module_args": {
                                    "dest": "~/.ssh/id_rsa",
                                    "force": false,
                                    "mode": 384,
                                    "src": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/a3240bd8b4c74e6f8c9446925cc2fed6_id_rsa"
                                },
                                "src": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/a3240bd8b4c74e6f8c9446925cc2fed6_id_rsa"
                            },
                            "src": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/a3240bd8b4c74e6f8c9446925cc2fed6_id_rsa"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:27.736344Z",
                            "start": "2026-03-21T00:00:27.134226Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-0000000000cd",
                        "name": "Install build private key as SSH key on all nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "copy",
                            "changed": false,
                            "dest": "/home/zuul-testbed04/.ssh/id_rsa.pub",
                            "invocation": {
                                "dest": "~/.ssh/id_rsa.pub",
                                "force": false,
                                "mode": 420,
                                "module_args": {
                                    "dest": "~/.ssh/id_rsa.pub",
                                    "force": false,
                                    "mode": 420,
                                    "src": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/a3240bd8b4c74e6f8c9446925cc2fed6_id_rsa.pub"
                                },
                                "src": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/a3240bd8b4c74e6f8c9446925cc2fed6_id_rsa.pub"
                            },
                            "src": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/a3240bd8b4c74e6f8c9446925cc2fed6_id_rsa.pub"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:28.334257Z",
                            "start": "2026-03-21T00:00:27.748561Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-0000000000ce",
                        "name": "Install build public key as SSH key on all nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "include_tasks",
                            "changed": false,
                            "false_condition": "ansible_os_family == \"Windows\"",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:28.413744Z",
                            "start": "2026-03-21T00:00:28.343157Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-0000000000ac",
                        "name": "Remote setup ssh keys (windows)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "sshagent_remove_keys",
                            "changed": true,
                            "invocation": {
                                "module_args": {
                                    "remove": "^(?!\\(stdin\\)).*"
                                }
                            },
                            "removed": [
                                "/var/ssh/nodepool"
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-0000000000b1",
                        "name": "remove-zuul-sshkey",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/remove-zuul-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:30.071688Z",
                            "start": "2026-03-21T00:00:28.419398Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-0000000000b3",
                        "name": "Remove master key from local agent"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": true,
                            "cmd": [
                                "ssh-add",
                                "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/a3240bd8b4c74e6f8c9446925cc2fed6_id_rsa"
                            ],
                            "delta": "0:00:00.025512",
                            "end": "2026-03-21 00:00:30.765952",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh-add /var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/a3240bd8b4c74e6f8c9446925cc2fed6_id_rsa",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-1af1-8b27-0000000000b9-0-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-21 00:00:30.740440",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "Identity added: /var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/a3240bd8b4c74e6f8c9446925cc2fed6_id_rsa (zuul-build-sshkey)",
                            "stdout_lines": [
                                "Identity added: /var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/a3240bd8b4c74e6f8c9446925cc2fed6_id_rsa (zuul-build-sshkey)"
                            ],
                            "zuul_log_id": "0242ac1f-6606-1af1-8b27-0000000000b9-0-orchestrator"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:30.812231Z",
                            "start": "2026-03-21T00:00:30.099279Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-0000000000b9",
                        "name": "Add back temp key"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ping",
                            "changed": false,
                            "invocation": {
                                "module_args": {
                                    "data": "pong"
                                }
                            },
                            "ping": "pong"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:31.424485Z",
                            "start": "2026-03-21T00:00:30.817976Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-0000000000ba",
                        "name": "Verify we can still SSH to all nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": false,
                            "false_condition": "ansible_os_family == \"Windows\"",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:31.482503Z",
                            "start": "2026-03-21T00:00:31.429118Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-0000000000bb",
                        "name": "Verify we can still SSH to all nodes (windows)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "zuul_console",
                            "changed": false,
                            "invocation": {
                                "module_args": {
                                    "path": "/tmp/console-{log_uuid}.log",
                                    "port": 19885,
                                    "state": "present"
                                }
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-00000000001b",
                        "name": "start-zuul-console",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/start-zuul-console"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:32.234651Z",
                            "start": "2026-03-21T00:00:31.586219Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-00000000001d",
                        "name": "Start zuul_console daemon."
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "zuul_info_dir": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/logs/zuul-info"
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:32.281683Z",
                            "start": "2026-03-21T00:00:32.252349Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-000000000021",
                        "name": "Define zuul_info_dir fact"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": false,
                            "diff": {
                                "after": {
                                    "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/logs/zuul-info"
                                },
                                "before": {
                                    "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/logs/zuul-info"
                                }
                            },
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": 493,
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/logs/zuul-info",
                                    "recurse": false,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0755",
                            "owner": "root",
                            "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/logs/zuul-info",
                            "size": 4096,
                            "state": "directory",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:33.272598Z",
                            "start": "2026-03-21T00:00:32.297164Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-000000000022",
                        "name": "Ensure Zuul Ansible directory exists"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "setup",
                            "ansible_facts": {
                                "ansible_all_ipv4_addresses": [
                                    "10.128.0.21"
                                ],
                                "ansible_all_ipv6_addresses": [
                                    "2a13:1a81:8000:2::33d",
                                    "fe80::f816:3eff:febf:60f0"
                                ],
                                "ansible_apparmor": {
                                    "status": "enabled"
                                },
                                "ansible_architecture": "x86_64",
                                "ansible_bios_date": "04/01/2014",
                                "ansible_bios_vendor": "SeaBIOS",
                                "ansible_bios_version": "1.15.0-1",
                                "ansible_board_asset_tag": "NA",
                                "ansible_board_name": "NA",
                                "ansible_board_serial": "NA",
                                "ansible_board_vendor": "NA",
                                "ansible_board_version": "NA",
                                "ansible_chassis_asset_tag": "NA",
                                "ansible_chassis_serial": "NA",
                                "ansible_chassis_vendor": "QEMU",
                                "ansible_chassis_version": "pc-i440fx-6.2",
                                "ansible_cmdline": {
                                    "BOOT_IMAGE": "/boot/vmlinuz-6.1.0-23-cloud-amd64",
                                    "console": "ttyS0,115200",
                                    "consoleblank": "0",
                                    "earlyprintk": "ttyS0,115200",
                                    "ro": true,
                                    "root": "UUID=97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                },
                                "ansible_date_time": {
                                    "date": "2026-03-21",
                                    "day": "21",
                                    "epoch": "1774051234",
                                    "epoch_int": "1774051234",
                                    "hour": "00",
                                    "iso8601": "2026-03-21T00:00:34Z",
                                    "iso8601_basic": "20260321T000034044137",
                                    "iso8601_basic_short": "20260321T000034",
                                    "iso8601_micro": "2026-03-21T00:00:34.044137Z",
                                    "minute": "00",
                                    "month": "03",
                                    "second": "34",
                                    "time": "00:00:34",
                                    "tz": "UTC",
                                    "tz_dst": "UTC",
                                    "tz_offset": "+0000",
                                    "weekday": "Saturday",
                                    "weekday_number": "6",
                                    "weeknumber": "11",
                                    "year": "2026"
                                },
                                "ansible_default_ipv4": {
                                    "address": "10.128.0.21",
                                    "alias": "ens3",
                                    "broadcast": "",
                                    "gateway": "10.128.0.17",
                                    "interface": "ens3",
                                    "macaddress": "fa:16:3e:bf:60:f0",
                                    "mtu": 1500,
                                    "netmask": "255.255.255.240",
                                    "network": "10.128.0.16",
                                    "prefix": "28",
                                    "type": "ether"
                                },
                                "ansible_default_ipv6": {
                                    "address": "2a13:1a81:8000:2::33d",
                                    "gateway": "fe80::f816:3eff:fe98:d9f",
                                    "interface": "ens3",
                                    "macaddress": "fa:16:3e:bf:60:f0",
                                    "mtu": 1500,
                                    "prefix": "128",
                                    "scope": "global",
                                    "type": "ether"
                                },
                                "ansible_device_links": {
                                    "ids": {
                                        "sda": [
                                            "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c"
                                        ],
                                        "sda1": [
                                            "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part1"
                                        ],
                                        "sda14": [
                                            "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part14"
                                        ],
                                        "sda15": [
                                            "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part15"
                                        ]
                                    },
                                    "labels": {},
                                    "masters": {},
                                    "uuids": {
                                        "sda1": [
                                            "97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                        ],
                                        "sda15": [
                                            "1E24-44C9"
                                        ]
                                    }
                                },
                                "ansible_devices": {
                                    "loop0": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop1": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop2": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop3": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop4": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop5": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop6": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop7": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "sda": {
                                        "holders": [],
                                        "host": "SCSI storage controller: Red Hat, Inc. Virtio SCSI",
                                        "links": {
                                            "ids": [
                                                "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c"
                                            ],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": "QEMU HARDDISK",
                                        "partitions": {
                                            "sda1": {
                                                "holders": [],
                                                "links": {
                                                    "ids": [
                                                        "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part1"
                                                    ],
                                                    "labels": [],
                                                    "masters": [],
                                                    "uuids": [
                                                        "97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                                    ]
                                                },
                                                "sectors": "209453023",
                                                "sectorsize": 512,
                                                "size": "99.87 GB",
                                                "start": "262144",
                                                "uuid": "97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                            },
                                            "sda14": {
                                                "holders": [],
                                                "links": {
                                                    "ids": [
                                                        "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part14"
                                                    ],
                                                    "labels": [],
                                                    "masters": [],
                                                    "uuids": []
                                                },
                                                "sectors": "6144",
                                                "sectorsize": 512,
                                                "size": "3.00 MB",
                                                "start": "2048",
                                                "uuid": null
                                            },
                                            "sda15": {
                                                "holders": [],
                                                "links": {
                                                    "ids": [
                                                        "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part15"
                                                    ],
                                                    "labels": [],
                                                    "masters": [],
                                                    "uuids": [
                                                        "1E24-44C9"
                                                    ]
                                                },
                                                "sectors": "253952",
                                                "sectorsize": 512,
                                                "size": "124.00 MB",
                                                "start": "8192",
                                                "uuid": "1E24-44C9"
                                            }
                                        },
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "209715200",
                                        "sectorsize": "512",
                                        "size": "100.00 GB",
                                        "support_discard": "4096",
                                        "vendor": "QEMU",
                                        "virtual": 1
                                    }
                                },
                                "ansible_distribution": "Debian",
                                "ansible_distribution_file_parsed": true,
                                "ansible_distribution_file_path": "/etc/os-release",
                                "ansible_distribution_file_variety": "Debian",
                                "ansible_distribution_major_version": "12",
                                "ansible_distribution_minor_version": "13",
                                "ansible_distribution_release": "bookworm",
                                "ansible_distribution_version": "12.13",
                                "ansible_dns": {
                                    "nameservers": [
                                        "81.163.194.9",
                                        "81.163.194.10",
                                        "2a13:1a81::11:53",
                                        "2a13:1a81::12:53"
                                    ],
                                    "search": [
                                        "."
                                    ]
                                },
                                "ansible_domain": "",
                                "ansible_effective_group_id": 1005,
                                "ansible_effective_user_id": 1004,
                                "ansible_ens3": {
                                    "active": true,
                                    "device": "ens3",
                                    "features": {
                                        "esp_hw_offload": "off [fixed]",
                                        "esp_tx_csum_hw_offload": "off [fixed]",
                                        "fcoe_mtu": "off [fixed]",
                                        "generic_receive_offload": "on",
                                        "generic_segmentation_offload": "on",
                                        "highdma": "on [fixed]",
                                        "hsr_dup_offload": "off [fixed]",
                                        "hsr_fwd_offload": "off [fixed]",
                                        "hsr_tag_ins_offload": "off [fixed]",
                                        "hsr_tag_rm_offload": "off [fixed]",
                                        "hw_tc_offload": "off [fixed]",
                                        "l2_fwd_offload": "off [fixed]",
                                        "large_receive_offload": "off [fixed]",
                                        "loopback": "off [fixed]",
                                        "macsec_hw_offload": "off [fixed]",
                                        "netns_local": "off [fixed]",
                                        "ntuple_filters": "off [fixed]",
                                        "receive_hashing": "off [fixed]",
                                        "rx_all": "off [fixed]",
                                        "rx_checksumming": "on [fixed]",
                                        "rx_fcs": "off [fixed]",
                                        "rx_gro_hw": "on",
                                        "rx_gro_list": "off",
                                        "rx_udp_gro_forwarding": "off",
                                        "rx_udp_tunnel_port_offload": "off [fixed]",
                                        "rx_vlan_filter": "on [fixed]",
                                        "rx_vlan_offload": "off [fixed]",
                                        "rx_vlan_stag_filter": "off [fixed]",
                                        "rx_vlan_stag_hw_parse": "off [fixed]",
                                        "scatter_gather": "on",
                                        "tcp_segmentation_offload": "on",
                                        "tls_hw_record": "off [fixed]",
                                        "tls_hw_rx_offload": "off [fixed]",
                                        "tls_hw_tx_offload": "off [fixed]",
                                        "tx_checksum_fcoe_crc": "off [fixed]",
                                        "tx_checksum_ip_generic": "on",
                                        "tx_checksum_ipv4": "off [fixed]",
                                        "tx_checksum_ipv6": "off [fixed]",
                                        "tx_checksum_sctp": "off [fixed]",
                                        "tx_checksumming": "on",
                                        "tx_esp_segmentation": "off [fixed]",
                                        "tx_fcoe_segmentation": "off [fixed]",
                                        "tx_gre_csum_segmentation": "off [fixed]",
                                        "tx_gre_segmentation": "off [fixed]",
                                        "tx_gso_list": "off [fixed]",
                                        "tx_gso_partial": "off [fixed]",
                                        "tx_gso_robust": "on [fixed]",
                                        "tx_ipxip4_segmentation": "off [fixed]",
                                        "tx_ipxip6_segmentation": "off [fixed]",
                                        "tx_lockless": "off [fixed]",
                                        "tx_nocache_copy": "off",
                                        "tx_scatter_gather": "on",
                                        "tx_scatter_gather_fraglist": "off [fixed]",
                                        "tx_sctp_segmentation": "off [fixed]",
                                        "tx_tcp6_segmentation": "on",
                                        "tx_tcp_ecn_segmentation": "on",
                                        "tx_tcp_mangleid_segmentation": "off",
                                        "tx_tcp_segmentation": "on",
                                        "tx_tunnel_remcsum_segmentation": "off [fixed]",
                                        "tx_udp_segmentation": "off [fixed]",
                                        "tx_udp_tnl_csum_segmentation": "off [fixed]",
                                        "tx_udp_tnl_segmentation": "off [fixed]",
                                        "tx_vlan_offload": "off [fixed]",
                                        "tx_vlan_stag_hw_insert": "off [fixed]",
                                        "vlan_challenged": "off [fixed]"
                                    },
                                    "hw_timestamp_filters": [],
                                    "ipv4": {
                                        "address": "10.128.0.21",
                                        "broadcast": "",
                                        "netmask": "255.255.255.240",
                                        "network": "10.128.0.16",
                                        "prefix": "28"
                                    },
                                    "ipv6": [
                                        {
                                            "address": "2a13:1a81:8000:2::33d",
                                            "prefix": "128",
                                            "scope": "global"
                                        },
                                        {
                                            "address": "fe80::f816:3eff:febf:60f0",
                                            "prefix": "64",
                                            "scope": "link"
                                        }
                                    ],
                                    "macaddress": "fa:16:3e:bf:60:f0",
                                    "module": "virtio_net",
                                    "mtu": 1500,
                                    "pciid": "virtio1",
                                    "promisc": false,
                                    "speed": -1,
                                    "timestamping": [],
                                    "type": "ether"
                                },
                                "ansible_env": {
                                    "HOME": "/home/zuul-testbed04",
                                    "LANG": "C.UTF-8",
                                    "LOGNAME": "zuul-testbed04",
                                    "MOTD_SHOWN": "pam",
                                    "PATH": "/usr/local/bin:/usr/bin:/bin:/usr/games",
                                    "PWD": "/home/zuul-testbed04",
                                    "SHELL": "/bin/sh",
                                    "SSH_CLIENT": "81.163.193.69 43364 22",
                                    "SSH_CONNECTION": "81.163.193.69 43364 10.128.0.21 22",
                                    "USER": "zuul-testbed04",
                                    "XDG_RUNTIME_DIR": "/run/user/1004",
                                    "XDG_SESSION_CLASS": "user",
                                    "XDG_SESSION_ID": "22803",
                                    "XDG_SESSION_TYPE": "tty"
                                },
                                "ansible_fibre_channel_wwn": [],
                                "ansible_fips": false,
                                "ansible_form_factor": "Other",
                                "ansible_fqdn": "zuul-static-regiocloud-infra-1",
                                "ansible_hostname": "zuul-static-regiocloud-infra-1",
                                "ansible_hostnqn": "",
                                "ansible_interfaces": [
                                    "lo",
                                    "ens3"
                                ],
                                "ansible_is_chroot": false,
                                "ansible_iscsi_iqn": "",
                                "ansible_kernel": "6.1.0-23-cloud-amd64",
                                "ansible_kernel_version": "#1 SMP PREEMPT_DYNAMIC Debian 6.1.99-1 (2024-07-15)",
                                "ansible_lo": {
                                    "active": true,
                                    "device": "lo",
                                    "features": {
                                        "esp_hw_offload": "off [fixed]",
                                        "esp_tx_csum_hw_offload": "off [fixed]",
                                        "fcoe_mtu": "off [fixed]",
                                        "generic_receive_offload": "on",
                                        "generic_segmentation_offload": "on",
                                        "highdma": "on [fixed]",
                                        "hsr_dup_offload": "off [fixed]",
                                        "hsr_fwd_offload": "off [fixed]",
                                        "hsr_tag_ins_offload": "off [fixed]",
                                        "hsr_tag_rm_offload": "off [fixed]",
                                        "hw_tc_offload": "off [fixed]",
                                        "l2_fwd_offload": "off [fixed]",
                                        "large_receive_offload": "off [fixed]",
                                        "loopback": "on [fixed]",
                                        "macsec_hw_offload": "off [fixed]",
                                        "netns_local": "on [fixed]",
                                        "ntuple_filters": "off [fixed]",
                                        "receive_hashing": "off [fixed]",
                                        "rx_all": "off [fixed]",
                                        "rx_checksumming": "on [fixed]",
                                        "rx_fcs": "off [fixed]",
                                        "rx_gro_hw": "off [fixed]",
                                        "rx_gro_list": "off",
                                        "rx_udp_gro_forwarding": "off",
                                        "rx_udp_tunnel_port_offload": "off [fixed]",
                                        "rx_vlan_filter": "off [fixed]",
                                        "rx_vlan_offload": "off [fixed]",
                                        "rx_vlan_stag_filter": "off [fixed]",
                                        "rx_vlan_stag_hw_parse": "off [fixed]",
                                        "scatter_gather": "on",
                                        "tcp_segmentation_offload": "on",
                                        "tls_hw_record": "off [fixed]",
                                        "tls_hw_rx_offload": "off [fixed]",
                                        "tls_hw_tx_offload": "off [fixed]",
                                        "tx_checksum_fcoe_crc": "off [fixed]",
                                        "tx_checksum_ip_generic": "on [fixed]",
                                        "tx_checksum_ipv4": "off [fixed]",
                                        "tx_checksum_ipv6": "off [fixed]",
                                        "tx_checksum_sctp": "on [fixed]",
                                        "tx_checksumming": "on",
                                        "tx_esp_segmentation": "off [fixed]",
                                        "tx_fcoe_segmentation": "off [fixed]",
                                        "tx_gre_csum_segmentation": "off [fixed]",
                                        "tx_gre_segmentation": "off [fixed]",
                                        "tx_gso_list": "on",
                                        "tx_gso_partial": "off [fixed]",
                                        "tx_gso_robust": "off [fixed]",
                                        "tx_ipxip4_segmentation": "off [fixed]",
                                        "tx_ipxip6_segmentation": "off [fixed]",
                                        "tx_lockless": "on [fixed]",
                                        "tx_nocache_copy": "off [fixed]",
                                        "tx_scatter_gather": "on [fixed]",
                                        "tx_scatter_gather_fraglist": "on [fixed]",
                                        "tx_sctp_segmentation": "on",
                                        "tx_tcp6_segmentation": "on",
                                        "tx_tcp_ecn_segmentation": "on",
                                        "tx_tcp_mangleid_segmentation": "on",
                                        "tx_tcp_segmentation": "on",
                                        "tx_tunnel_remcsum_segmentation": "off [fixed]",
                                        "tx_udp_segmentation": "on",
                                        "tx_udp_tnl_csum_segmentation": "off [fixed]",
                                        "tx_udp_tnl_segmentation": "off [fixed]",
                                        "tx_vlan_offload": "off [fixed]",
                                        "tx_vlan_stag_hw_insert": "off [fixed]",
                                        "vlan_challenged": "on [fixed]"
                                    },
                                    "hw_timestamp_filters": [],
                                    "ipv4": {
                                        "address": "127.0.0.1",
                                        "broadcast": "",
                                        "netmask": "255.0.0.0",
                                        "network": "127.0.0.0",
                                        "prefix": "8"
                                    },
                                    "ipv6": [
                                        {
                                            "address": "::1",
                                            "prefix": "128",
                                            "scope": "host"
                                        }
                                    ],
                                    "mtu": 65536,
                                    "promisc": false,
                                    "timestamping": [],
                                    "type": "loopback"
                                },
                                "ansible_loadavg": {
                                    "15m": 0.02587890625,
                                    "1m": 0.34814453125,
                                    "5m": 0.08056640625
                                },
                                "ansible_local": {},
                                "ansible_locally_reachable_ips": {
                                    "ipv4": [
                                        "10.128.0.21",
                                        "127.0.0.0/8",
                                        "127.0.0.1"
                                    ],
                                    "ipv6": [
                                        "::1",
                                        "2a13:1a81:8000:2::33d",
                                        "fe80::f816:3eff:febf:60f0"
                                    ]
                                },
                                "ansible_lsb": {
                                    "codename": "bookworm",
                                    "description": "Debian GNU/Linux 12 (bookworm)",
                                    "id": "Debian",
                                    "major_release": "12",
                                    "release": "12"
                                },
                                "ansible_lvm": "N/A",
                                "ansible_machine": "x86_64",
                                "ansible_machine_id": "28f0fb0ddd3c48c38b3aa31603cb444d",
                                "ansible_memfree_mb": 559,
                                "ansible_memory_mb": {
                                    "nocache": {
                                        "free": 7226,
                                        "used": 726
                                    },
                                    "real": {
                                        "free": 559,
                                        "total": 7952,
                                        "used": 7393
                                    },
                                    "swap": {
                                        "cached": 0,
                                        "free": 0,
                                        "total": 0,
                                        "used": 0
                                    }
                                },
                                "ansible_memtotal_mb": 7952,
                                "ansible_mounts": [
                                    {
                                        "block_available": 17172862,
                                        "block_size": 4096,
                                        "block_total": 25759134,
                                        "block_used": 8586272,
                                        "device": "/dev/sda1",
                                        "fstype": "ext4",
                                        "inode_available": 5912068,
                                        "inode_total": 6545408,
                                        "inode_used": 633340,
                                        "mount": "/",
                                        "options": "rw,relatime,discard,errors=remount-ro",
                                        "size_available": 70340042752,
                                        "size_total": 105509412864,
                                        "uuid": "97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                    },
                                    {
                                        "block_available": 57419,
                                        "block_size": 2048,
                                        "block_total": 63339,
                                        "block_used": 5920,
                                        "device": "/dev/sda15",
                                        "fstype": "vfat",
                                        "inode_available": 0,
                                        "inode_total": 0,
                                        "inode_used": 0,
                                        "mount": "/boot/efi",
                                        "options": "rw,relatime,fmask=0022,dmask=0022,codepage=437,iocharset=ascii,shortname=mixed,utf8,errors=remount-ro",
                                        "size_available": 117594112,
                                        "size_total": 129718272,
                                        "uuid": "1E24-44C9"
                                    }
                                ],
                                "ansible_nodename": "zuul-static-regiocloud-infra-1",
                                "ansible_os_family": "Debian",
                                "ansible_pkg_mgr": "apt",
                                "ansible_proc_cmdline": {
                                    "BOOT_IMAGE": "/boot/vmlinuz-6.1.0-23-cloud-amd64",
                                    "console": [
                                        "tty0",
                                        "ttyS0,115200"
                                    ],
                                    "consoleblank": "0",
                                    "earlyprintk": "ttyS0,115200",
                                    "ro": true,
                                    "root": "UUID=97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                },
                                "ansible_processor": [
                                    "0",
                                    "GenuineIntel",
                                    "Intel Xeon Processor (Icelake)",
                                    "1",
                                    "GenuineIntel",
                                    "Intel Xeon Processor (Icelake)",
                                    "2",
                                    "GenuineIntel",
                                    "Intel Xeon Processor (Icelake)",
                                    "3",
                                    "GenuineIntel",
                                    "Intel Xeon Processor (Icelake)"
                                ],
                                "ansible_processor_cores": 1,
                                "ansible_processor_count": 4,
                                "ansible_processor_nproc": 4,
                                "ansible_processor_threads_per_core": 1,
                                "ansible_processor_vcpus": 4,
                                "ansible_product_name": "OpenStack Nova",
                                "ansible_product_serial": "NA",
                                "ansible_product_uuid": "NA",
                                "ansible_product_version": "26.2.1",
                                "ansible_python": {
                                    "executable": "/usr/bin/python3",
                                    "has_sslcontext": true,
                                    "type": "cpython",
                                    "version": {
                                        "major": 3,
                                        "micro": 2,
                                        "minor": 11,
                                        "releaselevel": "final",
                                        "serial": 0
                                    },
                                    "version_info": [
                                        3,
                                        11,
                                        2,
                                        "final",
                                        0
                                    ]
                                },
                                "ansible_python_version": "3.11.2",
                                "ansible_real_group_id": 1005,
                                "ansible_real_user_id": 1004,
                                "ansible_selinux": {
                                    "status": "disabled"
                                },
                                "ansible_selinux_python_present": true,
                                "ansible_service_mgr": "systemd",
                                "ansible_ssh_host_key_dsa_public": "AAAAB3NzaC1kc3MAAACBAOVugYBPOvEOPI9DA30PvhAEr88mJeaFmKN4pkfnBHV/1A1OlDKtX7lKWCBfEUilcd+tQFVKXm4KY+K9q4J5x5tpd7cL7LRDCFKN+py3+jc6U5J9EZ0o5cVG84z0c3XLEL6cCljG9Etx/0pXjjELRh1SGuUZweEx0143OSiKs/9DAAAAFQCT7R5MDTPEdFXQcZxA/fPRmtWFswAAAIBp8Iz2Njfcxfnh/mdxMztzkdNXulDt679+ghvl5588eCIDPa9xMtojivk53JaV6V4R+wsBN+MaKufzMOoVxuDkZcmw940H/1kcr3yOsJcZIkU1GRkJiNnSEIWJ6dVaXMh+wwbjw+8xNCA3+n3fARqUK906ICT5RsRrBe7W2K7dIAAAAIAL9qFr0vTidcXXhyWKN8I2iBsHSKfb6cjbseShw3ZeZE+kJ3IXz3z9oe0OwKp52GhvJrOXEkxST9YbtutbsVz5iS7lDpisvIWfFtvRRYaOtyvE8b4PWpHrk4s5tssqGDQ7y7zSqmRrj1sf5w0pwokPGmoSnPWUi0Vp0yhlysI/1w==",
                                "ansible_ssh_host_key_dsa_public_keytype": "ssh-dss",
                                "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAuyO5R2SR4XT/Y65sItV7OQrUJXmGJ+//L9L/VugDohOWInFrkaZ6yFJMccVS7fmApekshaCGYnkIVdXV72hzg=",
                                "ansible_ssh_host_key_ecdsa_public_keytype": "ecdsa-sha2-nistp256",
                                "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAILxLVCAcH99lPquKCT3PwHsvi+569sV28I73LxchEjeg",
                                "ansible_ssh_host_key_ed25519_public_keytype": "ssh-ed25519",
                                "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABgQCZ3YX5Qhzq+HT8jTGGcX7rLsF3GCwObHFepfMmBZuD+gTfvdY5sYFSoXiT70rLPBjqETBTS0SQwmo8sd5AJI1vdgIqI5Sry8p52ng8tDg6C5CONt+xAaYOmE+ZDD/6naNX56b/gmgqBRQqPQHPmuptNxTlimdAcpoi+p9/UFBLTObMYmNL+5DTqw1+XECJ5j5kj5Ez40NT5w/1XnelEMcdw06WJZrzNEFt4Qpr2Zf+ybh0s+Xcn9u7d/gCbwXIPJvJnT0iSkiOzFvXJgOzwpvPtoi/5cV7g1LmgWsMLadT6gHof2poQaofC+1Bi0KiB2ANARMumvVBULx03qKvsqulsnlQAix2ffMraa3aMdUzK5aZo3hBx+H1uyEznUlyG9TUT5zhLIhNObetbpJgv6OJHiknBW6v/80FGD9t1pDiAwwBc/tJ8TyKpwyAdGMiEBIIVxTzx8QfPuApPJBJZXhMo2FBwRAEDEuuHbZu0znjhKKz4zlFT8TDrW+W7iThTvc=",
                                "ansible_ssh_host_key_rsa_public_keytype": "ssh-rsa",
                                "ansible_swapfree_mb": 0,
                                "ansible_swaptotal_mb": 0,
                                "ansible_system": "Linux",
                                "ansible_system_capabilities": [
                                    ""
                                ],
                                "ansible_system_capabilities_enforced": "True",
                                "ansible_system_vendor": "OpenStack Foundation",
                                "ansible_uptime_seconds": 49381209,
                                "ansible_user_dir": "/home/zuul-testbed04",
                                "ansible_user_gecos": "",
                                "ansible_user_gid": 1005,
                                "ansible_user_id": "zuul-testbed04",
                                "ansible_user_shell": "/bin/sh",
                                "ansible_user_uid": 1004,
                                "ansible_userspace_architecture": "x86_64",
                                "ansible_userspace_bits": "64",
                                "ansible_virtualization_role": "guest",
                                "ansible_virtualization_tech_guest": [
                                    "openstack"
                                ],
                                "ansible_virtualization_tech_host": [
                                    "kvm"
                                ],
                                "ansible_virtualization_type": "openstack",
                                "gather_subset": [
                                    "all"
                                ],
                                "module_setup": true
                            },
                            "changed": false,
                            "invocation": {
                                "module_args": {
                                    "fact_path": "/etc/ansible/facts.d",
                                    "filter": [],
                                    "gather_subset": [
                                        "all"
                                    ],
                                    "gather_timeout": 10
                                }
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:34.527804Z",
                            "start": "2026-03-21T00:00:33.278297Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-000000000023",
                        "name": "Collect information about the host"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "validate_host_hostname_sanitized": "orchestrator"
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:34.597128Z",
                            "start": "2026-03-21T00:00:34.545910Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-000000000024",
                        "name": "Sanitize hostname"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "template",
                            "changed": true,
                            "checksum": "e531388b8fbbb0d4b2890044f73a9dfe7b631613",
                            "dest": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/logs/zuul-info/host-info.orchestrator.yaml",
                            "diff": [],
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_original_basename": "host-info.j2",
                                    "attributes": null,
                                    "backup": false,
                                    "checksum": "e531388b8fbbb0d4b2890044f73a9dfe7b631613",
                                    "content": null,
                                    "dest": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/logs/zuul-info/host-info.orchestrator.yaml",
                                    "directory_mode": null,
                                    "follow": false,
                                    "force": true,
                                    "group": null,
                                    "local_follow": null,
                                    "mode": 420,
                                    "owner": null,
                                    "remote_src": null,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/.ansible/tmp/ansible-tmp-1774051234.6977904-238-188256465714154/source",
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "md5sum": "e500bcbf81fc1c113a9f52f8b0cdeaf4",
                            "mode": "0644",
                            "owner": "root",
                            "size": 19734,
                            "src": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/.ansible/tmp/ansible-tmp-1774051234.6977904-238-188256465714154/source",
                            "state": "file",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:35.754010Z",
                            "start": "2026-03-21T00:00:34.601756Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-000000000025",
                        "name": "Write out all ansible variables/facts known for each host"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "_zuul_nolog_return": true,
                            "action": "zuul_debug_info",
                            "changed": false,
                            "df_i": "Filesystem      Inodes  IUsed   IFree IUse% Mounted on\nudev           1015018    312 1014706    1% /dev\ntmpfs          1017896    542 1017354    1% /run\n/dev/sda1      6545408 633340 5912068   10% /\ntmpfs          1017896      1 1017895    1% /dev/shm\ntmpfs          1017896      2 1017894    1% /run/lock\n/dev/sda15           0      0       0     - /boot/efi\ntmpfs           203579     17  203562    1% /run/user/1001\ntmpfs           203579     17  203562    1% /run/user/1002\ntmpfs           203579     17  203562    1% /run/user/1006\ntmpfs           203579     17  203562    1% /run/user/1004\ntmpfs           203579     17  203562    1% /run/user/1005\n",
                            "df_m": "Filesystem     1M-blocks  Used Available Use% Mounted on\nudev                3965     0      3965   0% /dev\ntmpfs                796     1       795   1% /run\n/dev/sda1         100622 29415     67082  31% /\ntmpfs               3977     0      3977   0% /dev/shm\ntmpfs                  5     0         5   0% /run/lock\n/dev/sda15           124    12       113  10% /boot/efi\ntmpfs                796     0       796   0% /run/user/1001\ntmpfs                796     0       796   0% /run/user/1002\ntmpfs                796     0       796   0% /run/user/1006\ntmpfs                796     0       796   0% /run/user/1004\ntmpfs                796     0       796   0% /run/user/1005\n",
                            "image_manifest_files": [],
                            "invocation": {
                                "module_args": {
                                    "image_manifest": null,
                                    "image_manifest_files": [
                                        "/etc/dib-builddate.txt",
                                        "/etc/image-hostname.txt"
                                    ],
                                    "ipv4_route_required": false,
                                    "ipv6_route_required": false,
                                    "traceroute_host": null
                                }
                            },
                            "network_interfaces": "1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000\n    link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00\n    inet 127.0.0.1/8 scope host lo\n       valid_lft forever preferred_lft forever\n    inet6 ::1/128 scope host noprefixroute \n       valid_lft forever preferred_lft forever\n2: ens3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc fq_codel state UP group default qlen 1000\n    link/ether fa:16:3e:bf:60:f0 brd ff:ff:ff:ff:ff:ff\n    altname enp0s3\n    inet 10.128.0.21/28 metric 100 brd 10.128.0.31 scope global dynamic ens3\n       valid_lft 24830sec preferred_lft 24830sec\n    inet6 2a13:1a81:8000:2::33d/128 scope global noprefixroute \n       valid_lft forever preferred_lft forever\n    inet6 fe80::f816:3eff:febf:60f0/64 scope link \n       valid_lft forever preferred_lft forever\n",
                            "network_neighbors": "10.128.0.17 dev ens3 lladdr fa:16:3e:e3:0c:52 REACHABLE \nfe80::f816:3eff:fe98:d9f dev ens3 lladdr fa:16:3e:98:0d:9f router STALE \n",
                            "network_routing_v4": "default via 10.128.0.17 dev ens3 proto dhcp src 10.128.0.21 metric 100 \n10.128.0.16/28 dev ens3 proto kernel scope link src 10.128.0.21 metric 100 \n10.128.0.17 dev ens3 proto dhcp scope link src 10.128.0.21 metric 100 \n10.128.0.18 dev ens3 proto dhcp scope link src 10.128.0.21 metric 100 \n81.163.194.9 via 10.128.0.17 dev ens3 proto dhcp src 10.128.0.21 metric 100 \n81.163.194.10 via 10.128.0.17 dev ens3 proto dhcp src 10.128.0.21 metric 100 \n169.254.169.254 via 10.128.0.18 dev ens3 proto dhcp src 10.128.0.21 metric 100 \n",
                            "network_routing_v6": "2a13:1a81:8000:2::300/120 dev ens3 proto ra metric 100 pref medium\nfe80::/64 dev ens3 proto kernel metric 256 pref medium\ndefault via fe80::f816:3eff:fe98:d9f dev ens3 proto ra metric 100 mtu 1500 pref medium\n",
                            "proc_cpuinfo": "processor\t: 0\nvendor_id\t: GenuineIntel\ncpu family\t: 6\nmodel\t\t: 134\nmodel name\t: Intel Xeon Processor (Icelake)\nstepping\t: 0\nmicrocode\t: 0x1\ncpu MHz\t\t: 2299.998\ncache size\t: 16384 KB\nphysical id\t: 0\nsiblings\t: 1\ncore id\t\t: 0\ncpu cores\t: 1\napicid\t\t: 0\ninitial apicid\t: 0\nfpu\t\t: yes\nfpu_exception\t: yes\ncpuid level\t: 13\nwp\t\t: yes\nflags\t\t: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology cpuid tsc_known_freq pni pclmulqdq vmx ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch cpuid_fault invpcid_single ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves wbnoinvd arat avx512vbmi umip pku ospke avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq la57 rdpid fsrm md_clear arch_capabilities\nvmx flags\t: vnmi preemption_timer posted_intr invvpid ept_x_only ept_ad ept_1gb flexpriority apicv tsc_offset vtpr mtf vapic ept vpid unrestricted_guest vapic_reg vid shadow_vmcs\nbugs\t\t: spectre_v1 spectre_v2 spec_store_bypass swapgs taa mmio_stale_data rfds bhi\nbogomips\t: 4599.99\nclflush size\t: 64\ncache_alignment\t: 64\naddress sizes\t: 40 bits physical, 57 bits virtual\npower management:\n\nprocessor\t: 1\nvendor_id\t: GenuineIntel\ncpu family\t: 6\nmodel\t\t: 134\nmodel name\t: Intel Xeon Processor (Icelake)\nstepping\t: 0\nmicrocode\t: 0x1\ncpu MHz\t\t: 2299.998\ncache size\t: 16384 KB\nphysical id\t: 1\nsiblings\t: 1\ncore id\t\t: 0\ncpu cores\t: 1\napicid\t\t: 1\ninitial apicid\t: 1\nfpu\t\t: yes\nfpu_exception\t: yes\ncpuid level\t: 13\nwp\t\t: yes\nflags\t\t: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology cpuid tsc_known_freq pni pclmulqdq vmx ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch cpuid_fault invpcid_single ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves wbnoinvd arat avx512vbmi umip pku ospke avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq la57 rdpid fsrm md_clear arch_capabilities\nvmx flags\t: vnmi preemption_timer posted_intr invvpid ept_x_only ept_ad ept_1gb flexpriority apicv tsc_offset vtpr mtf vapic ept vpid unrestricted_guest vapic_reg vid shadow_vmcs\nbugs\t\t: spectre_v1 spectre_v2 spec_store_bypass swapgs taa mmio_stale_data rfds bhi\nbogomips\t: 4599.99\nclflush size\t: 64\ncache_alignment\t: 64\naddress sizes\t: 40 bits physical, 57 bits virtual\npower management:\n\nprocessor\t: 2\nvendor_id\t: GenuineIntel\ncpu family\t: 6\nmodel\t\t: 134\nmodel name\t: Intel Xeon Processor (Icelake)\nstepping\t: 0\nmicrocode\t: 0x1\ncpu MHz\t\t: 2299.998\ncache size\t: 16384 KB\nphysical id\t: 2\nsiblings\t: 1\ncore id\t\t: 0\ncpu cores\t: 1\napicid\t\t: 2\ninitial apicid\t: 2\nfpu\t\t: yes\nfpu_exception\t: yes\ncpuid level\t: 13\nwp\t\t: yes\nflags\t\t: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology cpuid tsc_known_freq pni pclmulqdq vmx ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch cpuid_fault invpcid_single ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves wbnoinvd arat avx512vbmi umip pku ospke avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq la57 rdpid fsrm md_clear arch_capabilities\nvmx flags\t: vnmi preemption_timer posted_intr invvpid ept_x_only ept_ad ept_1gb flexpriority apicv tsc_offset vtpr mtf vapic ept vpid unrestricted_guest vapic_reg vid shadow_vmcs\nbugs\t\t: spectre_v1 spectre_v2 spec_store_bypass swapgs taa mmio_stale_data rfds bhi\nbogomips\t: 4599.99\nclflush size\t: 64\ncache_alignment\t: 64\naddress sizes\t: 40 bits physical, 57 bits virtual\npower management:\n\nprocessor\t: 3\nvendor_id\t: GenuineIntel\ncpu family\t: 6\nmodel\t\t: 134\nmodel name\t: Intel Xeon Processor (Icelake)\nstepping\t: 0\nmicrocode\t: 0x1\ncpu MHz\t\t: 2299.998\ncache size\t: 16384 KB\nphysical id\t: 3\nsiblings\t: 1\ncore id\t\t: 0\ncpu cores\t: 1\napicid\t\t: 3\ninitial apicid\t: 3\nfpu\t\t: yes\nfpu_exception\t: yes\ncpuid level\t: 13\nwp\t\t: yes\nflags\t\t: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology cpuid tsc_known_freq pni pclmulqdq vmx ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch cpuid_fault invpcid_single ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves wbnoinvd arat avx512vbmi umip pku ospke avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq la57 rdpid fsrm md_clear arch_capabilities\nvmx flags\t: vnmi preemption_timer posted_intr invvpid ept_x_only ept_ad ept_1gb flexpriority apicv tsc_offset vtpr mtf vapic ept vpid unrestricted_guest vapic_reg vid shadow_vmcs\nbugs\t\t: spectre_v1 spectre_v2 spec_store_bypass swapgs taa mmio_stale_data rfds bhi\nbogomips\t: 4599.99\nclflush size\t: 64\ncache_alignment\t: 64\naddress sizes\t: 40 bits physical, 57 bits virtual\npower management:\n\n",
                            "traceroute": null,
                            "uname": "Linux zuul-static-regiocloud-infra-1 6.1.0-23-cloud-amd64 #1 SMP PREEMPT_DYNAMIC Debian 6.1.99-1 (2024-07-15) x86_64 GNU/Linux\n"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:36.318725Z",
                            "start": "2026-03-21T00:00:35.758870Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-000000000027",
                        "name": "Collect information about zuul worker"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "template",
                            "changed": true,
                            "checksum": "cafe40b442e0b0edacdd34e339adb0195d98f2ba",
                            "dest": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/logs/zuul-info/zuul-info.orchestrator.txt",
                            "diff": [],
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_original_basename": "zuul-info.j2",
                                    "attributes": null,
                                    "backup": false,
                                    "checksum": "cafe40b442e0b0edacdd34e339adb0195d98f2ba",
                                    "content": null,
                                    "dest": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/logs/zuul-info/zuul-info.orchestrator.txt",
                                    "directory_mode": null,
                                    "follow": false,
                                    "force": true,
                                    "group": null,
                                    "local_follow": null,
                                    "mode": 420,
                                    "owner": null,
                                    "remote_src": null,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/.ansible/tmp/ansible-tmp-1774051236.4153433-277-114908713666985/source",
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "md5sum": "543f844ae49bb28ad6d51f5f05ae210a",
                            "mode": "0644",
                            "owner": "root",
                            "size": 3489,
                            "src": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/.ansible/tmp/ansible-tmp-1774051236.4153433-277-114908713666985/source",
                            "state": "file",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:37.780571Z",
                            "start": "2026-03-21T00:00:36.327192Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-000000000028",
                        "name": "Write out all zuul information for each host"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "zuul_console",
                            "changed": false,
                            "invocation": {
                                "module_args": {
                                    "path": "/tmp/console-{log_uuid}.log",
                                    "port": 19885,
                                    "state": "present"
                                }
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-00000000002a",
                        "name": "prepare-workspace-log",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_0/github.com/osism/zuul-config/roles/prepare-workspace-log"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:00:38.158231Z",
                            "start": "2026-03-21T00:00:37.793568Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-00000000002c",
                        "name": "Start zuul_console daemon."
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "synchronize",
                            "changed": true,
                            "cmd": "/usr/bin/rsync --compress --delete-after --archive --no-owner --no-group --rsh='/usr/bin/ssh -S none -o Port=22 -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null' --out-format='<<CHANGED>>%i %n%L' /var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/src zuul-testbed04@81.163.193.140:.",
                            "invocation": {
                                "module_args": {
                                    "_local_rsync_password": null,
                                    "_local_rsync_path": "rsync",
                                    "_ssh_args": null,
                                    "_substitute_controller": false,
                                    "archive": true,
                                    "checksum": false,
                                    "compress": true,
                                    "copy_links": false,
                                    "delay_updates": false,
                                    "delete": true,
                                    "dest": "zuul-testbed04@81.163.193.140:.",
                                    "dest_port": 22,
                                    "dirs": false,
                                    "existing_only": false,
                                    "group": false,
                                    "link_dest": null,
                                    "links": null,
                                    "mode": "push",
                                    "owner": false,
                                    "partial": false,
                                    "perms": null,
                                    "private_key": null,
                                    "recursive": true,
                                    "rsync_opts": [],
                                    "rsync_path": null,
                                    "rsync_timeout": 0,
                                    "set_remote_user": true,
                                    "src": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/src",
                                    "ssh_connection_multiplexing": false,
                                    "times": null,
                                    "use_ssh_args": false,
                                    "verify_host": false
                                }
                            },
                            "msg": ".d..t...... src/\n.d..t...... src/github.com/\n.d..t...... src/github.com/osism/\n.d..t...... src/github.com/osism/ansible-collection-commons/\n<f..t...... src/github.com/osism/ansible-collection-commons/.ansible-lint\n<f..t...... src/github.com/osism/ansible-collection-commons/.flake8\n<f..t...... src/github.com/osism/ansible-collection-commons/.gitignore\n<f..t...... src/github.com/osism/ansible-collection-commons/.yamllint.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/.zuul.yaml\n<f..t...... src/github.com/osism/ansible-collection-commons/CHANGELOG.md\n<f..t...... src/github.com/osism/ansible-collection-commons/LICENSE\n<f..t...... src/github.com/osism/ansible-collection-commons/README.md\n<f..t...... src/github.com/osism/ansible-collection-commons/galaxy.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/HEAD\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/ORIG_HEAD\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/config\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/config.worktree\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/description\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/index\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/packed-refs\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/branches/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/applypatch-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/commit-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/fsmonitor-watchman.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/post-update.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-applypatch.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-commit.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-merge-commit.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-push.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-rebase.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-receive.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/prepare-commit-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/push-to-checkout.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/update.sample\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/info/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/info/exclude\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/HEAD\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/heads/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/heads/main\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/origin/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/origin/HEAD\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/01/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/02/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/04/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/05/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/06/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/07/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/08/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/09/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/10/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/11/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/12/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/13/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/14/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/15/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/16/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/17/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/18/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/19/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/20/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/21/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/22/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/23/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/24/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/25/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/26/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/27/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/28/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/29/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/30/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/31/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/32/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/33/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/35/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/36/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/37/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/38/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/39/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/40/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/41/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/42/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/43/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/44/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/45/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/46/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/47/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/48/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/49/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/50/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/52/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/53/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/54/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/55/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/56/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/57/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/58/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/59/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/60/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/61/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/62/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/63/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/64/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/66/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/67/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/68/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/69/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/70/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/71/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/72/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/73/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/74/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/75/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/76/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/77/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/78/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/79/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/80/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/81/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/82/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/83/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/84/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/85/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/86/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/87/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/88/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/89/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/90/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/91/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/93/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/94/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/95/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/96/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/97/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/98/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a1/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a8/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/aa/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ab/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ac/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ad/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ae/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/af/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b1/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b8/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ba/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bb/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bc/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bd/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/be/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bf/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ca/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cb/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cc/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cd/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ce/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cf/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d1/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d8/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/da/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/db/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/dc/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/de/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/df/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e8/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ea/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/eb/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ec/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ed/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ee/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ef/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f1/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f8/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fa/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fb/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fc/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fd/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fe/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ff/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/info/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/pack/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/refs/\n.d..t...... src/github.com/osism/ansible-collection-commons/.github/\n<f..t...... src/github.com/osism/ansible-collection-commons/.github/publish-collection.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/.github/renovate.json\n.d..t...... src/github.com/osism/ansible-collection-commons/.github/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/.github/templates/galaxy.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/.github/workflows/\n<f..t...... src/github.com/osism/ansible-collection-commons/.github/workflows/publish-collection.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/meta/runtime.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/requirements.txt\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/collections.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/converge.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/molecule.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/certificates.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/cleanup.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/configfs.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/configuration.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/docker_compose.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/docker_login.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/facts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/firewall.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/hostname.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/ipmitool.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/k9s.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/kernel_modules.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/known_hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/limits.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/lynis.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/microcode.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/motd.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/network.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/operator.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/packages.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/podman.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/proxy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/repository.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/resolvconf.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/runc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/services.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sosreport.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sshconfig.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/state.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sysctl.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/systohc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/timezone.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/trivy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/user.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/certificates.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configfs.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/docker_compose.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/docker_login.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/facts.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/firewall.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hostname.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/ipmitool.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/k9s.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/kernel_modules.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/known_hosts.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/limits.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/microcode.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/operator.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/resolvconf.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/runc.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/services.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sosreport.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sshconfig.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/state.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sysctl.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/systohc.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/user.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/cloudinit.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/packages.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/packages_debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/services.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/timer_debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/timer_redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/git.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/main.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/type-block.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/type-template.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/interfaces.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/netplan.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/main.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/centos.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/debian.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/util.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/certificates.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/cleanup.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/configfs.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/configuration.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/docker_compose.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/docker_login.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/facts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/firewall.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/hostname.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/ipmitool.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/k9s.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/kernel_modules.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/known_hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/limits.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/lynis.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/microcode.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/motd.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/network.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/operator.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/packages.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/podman.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/proxy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/repository.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/resolvconf.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/runc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/services.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sosreport.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sshconfig.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/state.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sysctl.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/systohc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/timezone.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/trivy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/user.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/certificates.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/cleanup.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/configfs.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/configuration.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/docker_compose.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/facts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/firewall.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/hostname.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/ipmitool.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/k9s.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/kernel_modules.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/lynis.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/microcode.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/motd.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/operator.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/packages.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/podman.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/proxy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/repository.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/resolvconf.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/runc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/services.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sosreport.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sshconfig.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/state.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sysctl.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/systohc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/timezone.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/trivy.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/plugins/\n.d..t...... src/github.com/osism/ansible-collection-commons/plugins/callback/\n<f..t...... src/github.com/osism/ansible-collection-commons/plugins/callback/still_alive.py\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/templates/certificates.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/cloudinit.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/packages-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/packages-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/services-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/services-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/timers-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/timers-RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/git.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/update.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config-git-netbox-private-key.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config-git-private-key.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/files/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/files/docker-compose\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/docker-compose.service.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/osism.target.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/templates/ufw.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/templates/config.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-block.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-fast.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-local.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-template.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/hosts-block.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/hosts-fast.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/hosts-template.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/templates/module-load.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/write-scanned.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/write-static.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/limits.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/configure-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/configure-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/templates/motd.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-interfaces.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-netplan.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-networkd.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/dummy-interfaces.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/interfaces-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/interfaces-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/netplan-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/netplan-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/network-extra-init.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/vxlan-interfaces.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/dummy.netdev.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/network-extra-init.service.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/network-extra-init.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/vxlan.netdev.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/vxlan.network.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/device.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/interfaces.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/01-osism.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/interfaces.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/interfaces.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/templates/operator-sudoers.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/package-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/package-RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/deploy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/templates/proxy-apt.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/\n.L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/AlmaLinux.yml -> RedHat.yml\n.L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/CentOS.yml -> RedHat.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Debian.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/RedHat.yml\n.L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Rocky.yml -> RedHat.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Ubuntu.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/99osism.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/source.list.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/ubuntu.sources.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/AlmaLinux-amd64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/CentOS-x86_64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Debian-x86_64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Rocky-x86_64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Ubuntu-aarch64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Ubuntu-x86_64.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-resolv.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/templates/resolved.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/services/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/services/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/services/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/services/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/CentOS-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/Debian-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/Ubuntu-dist.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/templates/config.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/state/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/state/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/state/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/state/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/sysctl.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/remote-key.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/type-default.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/type-keyfile.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/templates/config.j2\n.d..t...... src/github.com/osism/ansible-collection-services/\n<f..t...... src/github.com/osism/ansible-collection-services/.ansible-lint\n<f..t...... src/github.com/osism/ansible-collection-services/.flake8\n<f..t...... src/github.com/osism/ansible-collection-services/.gitignore\n<f..t...... src/github.com/osism/ansible-collection-services/.yamllint.yml\n<f..t...... src/github.com/osism/ansible-collection-services/.zuul.yaml\n<f..t...... src/github.com/osism/ansible-collection-services/CHANGELOG.md\n<f..t...... src/github.com/osism/ansible-collection-services/LICENSE\n<f..t...... src/github.com/osism/ansible-collection-services/README.md\n<f..t...... src/github.com/osism/ansible-collection-services/galaxy.yml\n.d..t...... src/github.com/osism/ansible-collection-services/.git/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/HEAD\n<f..t...... src/github.com/osism/ansible-collection-services/.git/ORIG_HEAD\n<f..t...... src/github.com/osism/ansible-collection-services/.git/config\n<f..t...... src/github.com/osism/ansible-collection-services/.git/config.worktree\n<f..t...... src/github.com/osism/ansible-collection-services/.git/description\n<f..t...... src/github.com/osism/ansible-collection-services/.git/index\n<f..t...... src/github.com/osism/ansible-collection-services/.git/packed-refs\n.d..t...... src/github.com/osism/ansible-collection-services/.git/branches/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/hooks/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/applypatch-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/commit-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/fsmonitor-watchman.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/post-update.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-applypatch.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-commit.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-merge-commit.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-push.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-rebase.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-receive.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/prepare-commit-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/push-to-checkout.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/update.sample\n.d..t...... src/github.com/osism/ansible-collection-services/.git/info/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/info/exclude\n.d..t...... src/github.com/osism/ansible-collection-services/.git/logs/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/HEAD\n.d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/heads/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/heads/main\n.d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/origin/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/origin/HEAD\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/00/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/01/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/02/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/03/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/05/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/07/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/08/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/09/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/10/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/11/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/12/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/13/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/14/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/16/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/19/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1d/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1e/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1f/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/20/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/21/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/22/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/23/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/24/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/25/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/26/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/27/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/28/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/29/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2d/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2e/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/30/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/31/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/32/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/33/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/34/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/38/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3e/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3f/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/40/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/41/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/42/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/43/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/44/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/45/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/49/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4e/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4f/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/50/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/51/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/53/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/54/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/55/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/56/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/57/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/58/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/59/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/5a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/5b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/60/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/61/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/63/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/64/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/65/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/66/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/67/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/68/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/69/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6f/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/70/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/71/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/72/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/73/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/74/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/75/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/76/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/77/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/79/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7d/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7e/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7f/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/80/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/81/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/82/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/83/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/84/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/85/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/87/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/88/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/89/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8d/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8e/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/91/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/92/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/93/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/94/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/95/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/96/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/97/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/98/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9f/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a0/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a1/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a2/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a3/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a4/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a5/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a8/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/aa/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ac/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ad/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ae/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/af/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b0/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b1/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b2/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b3/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b4/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b6/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b7/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b8/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b9/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ba/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/bb/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/bc/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/bd/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c0/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c2/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c3/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c4/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c5/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c6/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c9/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/cb/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/cc/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ce/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/cf/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d0/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d1/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d2/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d4/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d5/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d7/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d8/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d9/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/da/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/db/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/dc/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/dd/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/de/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/df/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e0/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e1/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e3/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e5/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e6/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e7/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e8/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e9/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ea/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/eb/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ec/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ed/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ee/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ef/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f0/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f2/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f3/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f6/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f7/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f8/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fa/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fd/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fe/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ff/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/info/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/pack/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/refs/\n.d..t...... src/github.com/osism/ansible-collection-services/.github/\n<f..t...... src/github.com/osism/ansible-collection-services/.github/publish-collection.yml\n<f..t...... src/github.com/osism/ansible-collection-services/.github/renovate.json\n.d..t...... src/github.com/osism/ansible-collection-services/.github/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/.github/templates/galaxy.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/.github/workflows/\n<f..t...... src/github.com/osism/ansible-collection-services/.github/workflows/publish-collection.yml\n.d..t...... src/github.com/osism/ansible-collection-services/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/meta/runtime.yml\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/requirements.txt\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/collections.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/converge.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/molecule.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify.yml\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/adminer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/auditd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/cephclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/cgit.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/chrony.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/clamav.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/containerd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/dnsdist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/dnsmasq.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/docker.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/fail2ban.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/falco.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/frr.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/gnmic.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/hddtemp.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/homer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/httpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/journald.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/lldpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/manager.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netbird.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netdata.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/nexus.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/openstackclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/opentelemetry_collector.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/osquery.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/phpmyadmin.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/rng.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/rsyslog.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/scaphandre.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/smartd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/squid.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/sshd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/stepca.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/substation.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/teleport.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/thanos_sidecar.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/traefik.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/tuned.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/wazuh_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/wireguard.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/zabbix_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/zuul.yml\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/adminer.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/auditd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cgit.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/dnsdist.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/dnsmasq.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/fail2ban.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/frr.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/gnmic.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/homer.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/httpd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/journald.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/lldpd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/manager.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbox.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/nexus.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/opentelemetry_collector.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/phpmyadmin.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/rng.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/rsyslog.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/scaphandre.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/smartd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/squid.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/sshd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/stepca.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/substation.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/thanos_sidecar.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/traefik.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/tuned.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wireguard.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zuul.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/container.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/package.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/package_debian.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/main.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/main.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/main.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/debian.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/container.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/package-debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/package-redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/util.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/adminer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/auditd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/cephclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/cgit.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/chrony.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/clamav.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/containerd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/dnsdist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/dnsmasq.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/docker.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/fail2ban.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/falco.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/frr.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/gnmic.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/hddtemp.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/homer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/httpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/journald.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/lldpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/manager.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netbird.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netdata.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/nexus.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/openstackclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/opentelemetry_collector.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/osquery.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/phpmyadmin.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/rng.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/rsyslog.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/scaphandre.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/smartd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/squid.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/sshd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/stepca.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/substation.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/teleport.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/thanos_sidecar.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/traefik.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/tuned.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/wazuh_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/wireguard.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/zabbix_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/zuul.yml\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/adminer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/auditd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/cephclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/cgit.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/chrony.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/clamav.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/containerd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/dnsdist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/dnsmasq.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/docker.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/fail2ban.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/falco.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/frr.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/gnmic.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/hddtemp.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/homer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/httpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/journald.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/lldpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/netbird.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/netdata.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/osquery.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/rng.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/rsyslog.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/smartd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/sshd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/teleport.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/tuned.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/wazuh_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/wireguard.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/zabbix_agent.yml\n.d..t...... src/github.com/osism/ansible-collection-services/plugins/\n.d..t...... src/github.com/osism/ansible-collection-services/plugins/filter/\n<f..t...... src/github.com/osism/ansible-collection-services/plugins/filter/address.py\n.d..t...... src/github.com/osism/ansible-collection-services/plugins/modules/\n<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_container_facts.py\n<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_docker.py\n<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_toolbox.py\n.d..t...... src/github.com/osism/ansible-collection-services/roles/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/auditd.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/rules/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/rules/20-neo23x0.rules.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-container.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-package-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-package-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/container.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/package-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/package-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/rook.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/ceph.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/keyring.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/ceph-authtool.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/ceph.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/crushtool.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rados.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/radosgw-admin.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rbd.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rook.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/cgitrc.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/httpd-cgit.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/templates/chrony.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/initialize.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/templates/config.toml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/CentOS-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/Debian-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/Ubuntu-dist.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/dnsdist.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/dnsmasq.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/.gitattributes\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/limits.conf\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/docker_containers.fact\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/docker_images.fact\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/bootstrap.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/facts.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/install-docker-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/install-docker-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/storage-block.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/storage-zram.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/config.json.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/containerd_config.toml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/daemon.json.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/overlay.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/templates/sshd.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/.gitignore\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/files/falco.service\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-RedHat-family_broken.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/daemons.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_k3s_cilium.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_leaf.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_loadbalancer.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_loadbalancer_external_uplink.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_test.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_yrzn001.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_yrzn_metalbox.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_yrzn_network.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/vtysh.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/gnmic.yaml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/RedHat-family/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/RedHat-family/hddtemp\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/service-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/service-RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/config.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/files/prepare-ironic-volume.sh\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/ironic.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/sonic-ztp.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/htaccess.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/httpd.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/post_install.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/ztp.json.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/templates/journald.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/config.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-ansible.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-ara.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-celery.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-frontend.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-scripts.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-vault.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-wrapper.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/initialize-vault.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/initialize.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/verify-versions.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/check-service-manager-versions.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-netbox-secret.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-ssh-key.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-vault-key.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/manager.service.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/vault.hcl.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/all.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ansible.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ara-server.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ara.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/client.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/conductor.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/frontend.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/inventory-reconciler.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/listener.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/mariadb.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/netbox.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/openstack.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/postgres.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/vault.env.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/scripts/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/scripts/osism-include.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/cilium.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/flux.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/hubble.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-ansible.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-ceph.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-custom.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-generic.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-infrastructure.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-kolla.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-manager.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-monitoring.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-openstack.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-run-without-secrets.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-run.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-state.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-update-docker.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-update-manager.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-validate.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/scripts.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/wrapper.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/startup-scripts/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/startup-scripts/270_tags.py\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config-netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config-postgres.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/rescue-service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/restart-service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/wait-for-healthy-service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-netbox-database.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-netbox-secret.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-postgres-secret.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/configuration.py.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/netbox.service.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/nginx-unit-1.26.json.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/nginx-unit.json.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/postgresql.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/netbox.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/postgres.env.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/custom_fields.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/device_roles.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/device_types.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/groups.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/manufacturers.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/object_permissions.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/prefix_vlan_roles.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/sites.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/tags.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/users.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/vars/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/client.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/server.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/netdata.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/stream.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/create_repos_from_list.groovy\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/setup_http_proxy.groovy\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/setup_realms.groovy\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/update_admin_password.groovy\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/call-script.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/declare-script.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/initialize.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.properties.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.service.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/container-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/container-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/package-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/package-RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/openstack.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/otel-collector-config.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rng/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rng/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rng/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/additional-log-server.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/fluentd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/logrotate.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/70-fluentd.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/71-additional-log-server.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/rsyslog-logrotate.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/rsyslog.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/CentOS-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/Debian-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/Ubuntu-dist.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/files/smartmontools\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/osism.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/config-wrapper.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/provisioner-acme.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/stepca.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/stepca.service.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/wrapper/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/wrapper/step.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/files/entrypoint.sh\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/substation.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/templates/teleport.yaml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/certificates.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/dynamic.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik-cert.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik-key.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/client.keys.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/ossec.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/client.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/server.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/templates/zabbix_agent2.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/builder-logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/executor-logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/launcher-logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/mime.types\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/openssl.cnf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/web-logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/zk-ca.sh\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/copy_tasks.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/generate_certs_and_keys.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/template_tasks.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/docker-compose.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/httpd.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/main.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/nodepool.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/wait-to-start-gearman.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/wait-to-start.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/zoo.cfg.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/zuul.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/executor.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/mariadb.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/nodepool_builder.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/nodepool_launcher.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/scheduler.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/web.env.j2\n.d..t...... src/github.com/osism/terraform-base/\n<f..t...... src/github.com/osism/terraform-base/.yamllint.yml\n<f..t...... src/github.com/osism/terraform-base/.zuul.yaml\n<f..t...... src/github.com/osism/terraform-base/LICENSE\n<f..t...... src/github.com/osism/terraform-base/README.md\n.d..t...... src/github.com/osism/terraform-base/.git/\n<f..t...... src/github.com/osism/terraform-base/.git/HEAD\n<f..t...... src/github.com/osism/terraform-base/.git/ORIG_HEAD\n<f..t...... src/github.com/osism/terraform-base/.git/config\n<f..t...... src/github.com/osism/terraform-base/.git/config.worktree\n<f..t...... src/github.com/osism/terraform-base/.git/description\n<f..t...... src/github.com/osism/terraform-base/.git/index\n<f..t...... src/github.com/osism/terraform-base/.git/packed-refs\n.d..t...... src/github.com/osism/terraform-base/.git/branches/\n.d..t...... src/github.com/osism/terraform-base/.git/hooks/\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/applypatch-msg.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/commit-msg.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/fsmonitor-watchman.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/post-update.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-applypatch.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-commit.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-merge-commit.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-push.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-rebase.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-receive.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/prepare-commit-msg.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/push-to-checkout.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/update.sample\n.d..t...... src/github.com/osism/terraform-base/.git/info/\n<f..t...... src/github.com/osism/terraform-base/.git/info/exclude\n.d..t...... src/github.com/osism/terraform-base/.git/logs/\n<f..t...... src/github.com/osism/terraform-base/.git/logs/HEAD\n.d..t...... src/github.com/osism/terraform-base/.git/logs/refs/\n.d..t...... src/github.com/osism/terraform-base/.git/logs/refs/heads/\n<f..t...... src/github.com/osism/terraform-base/.git/logs/refs/heads/main\n.d..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/\n.d..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/origin/\n<f..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/origin/HEAD\n.d..t...... src/github.com/osism/terraform-base/.git/objects/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/00/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/01/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/02/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/05/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/06/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/07/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/08/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/0a/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/0b/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/0d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/0f/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/10/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/12/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/15/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/16/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/17/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/1b/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/1c/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/1f/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/20/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/21/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/24/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/26/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/29/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/2c/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/36/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/37/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/3a/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/3d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/41/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/43/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/46/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/47/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/48/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/49/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/4b/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/4d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/52/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/55/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/56/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/58/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/5d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/61/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/63/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/66/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/67/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/68/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/6d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/70/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/72/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/76/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/7c/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/7f/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/80/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/81/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/82/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/83/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/84/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/85/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/86/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/88/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/8c/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/8d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/90/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/91/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/92/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/95/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/96/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/98/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/9b/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/9e/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a2/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a3/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a6/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a7/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a9/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/aa/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/af/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b0/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b3/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b4/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b8/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b9/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ba/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/bc/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/bf/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c0/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c1/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c4/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c7/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c8/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/cb/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ce/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/cf/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/d2/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/d3/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/d6/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/d7/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/d9/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/da/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/db/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/dc/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/dd/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/de/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/df/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/e1/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/e2/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/e3/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/e4/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/e8/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ea/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ee/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ef/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/f1/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/f2/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/f4/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/f9/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/fb/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ff/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/info/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/pack/\n.d..t...... src/github.com/osism/terraform-base/.git/refs/\n.d..t...... src/github.com/osism/terraform-base/.github/\n<f..t...... src/github.com/osism/terraform-base/.github/renovate.json\n.d..t...... src/github.com/osism/terraform-base/playbooks/\n<f..t...... src/github.com/osism/terraform-base/playbooks/validate.yml\n.d..t...... src/github.com/osism/terraform-base/testbed-default/\n<f..t...... src/github.com/osism/terraform-base/testbed-default/data.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/main.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/manager.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/neutron.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/nodes.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/nova.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/outputs.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/provider.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/variables.tf\n.d..t...... src/github.com/osism/terraform-base/testbed-default/customisations/\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_floatingip_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_ipv4_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_ipv6_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/default_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/external_api_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/neutron_floatingip_custom.tf\n.d..t...... src/github.com/osism/terraform-base/testbed-default/environments/\n<f..t...... src/github.com/osism/terraform-base/testbed-default/environments/default.tfvars\n.d..t...... src/github.com/osism/terraform-base/testbed-default/overrides/\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/manager_boot_from_image_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/manager_boot_from_volume_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_availability_zone_hints_network_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_availability_zone_hints_router_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_router_enable_snat_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_boot_from_image_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_boot_from_volume_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_use_ephemeral_storage_override.tf\n.d..t...... src/github.com/osism/testbed/\n<f..t...... src/github.com/osism/testbed/.ansible-lint\n<f..t...... src/github.com/osism/testbed/.flake8\n<f..t...... src/github.com/osism/testbed/.gitignore\n<f..t...... src/github.com/osism/testbed/.yamllint.yml\n<f..t...... src/github.com/osism/testbed/.zuul.yaml\n<f..t...... src/github.com/osism/testbed/LICENSE\n<f..t...... src/github.com/osism/testbed/Makefile\n<f..t...... src/github.com/osism/testbed/README.md\n<f..t...... src/github.com/osism/testbed/ansible.cfg\n<f..t...... src/github.com/osism/testbed/gilt.yml\n<f..t...... src/github.com/osism/testbed/requirements.txt\n<f..t...... src/github.com/osism/testbed/requirements.yml\n<f..t...... src/github.com/osism/testbed/tox.ini\n.d..t...... src/github.com/osism/testbed/.git/\n<f..t...... src/github.com/osism/testbed/.git/HEAD\n<f..t...... src/github.com/osism/testbed/.git/ORIG_HEAD\n<f..t...... src/github.com/osism/testbed/.git/config\n<f..t...... src/github.com/osism/testbed/.git/config.worktree\n<f..t...... src/github.com/osism/testbed/.git/description\n<f..t...... src/github.com/osism/testbed/.git/index\n<f..t...... src/github.com/osism/testbed/.git/packed-refs\n.d..t...... src/github.com/osism/testbed/.git/branches/\n.d..t...... src/github.com/osism/testbed/.git/hooks/\n<f..t...... src/github.com/osism/testbed/.git/hooks/applypatch-msg.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/commit-msg.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/fsmonitor-watchman.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/post-update.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-applypatch.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-commit.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-merge-commit.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-push.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-rebase.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-receive.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/prepare-commit-msg.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/push-to-checkout.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/update.sample\n.d..t...... src/github.com/osism/testbed/.git/info/\n<f..t...... src/github.com/osism/testbed/.git/info/exclude\n.d..t...... src/github.com/osism/testbed/.git/logs/\n<f..t...... src/github.com/osism/testbed/.git/logs/HEAD\n.d..t...... src/github.com/osism/testbed/.git/logs/refs/\n.d..t...... src/github.com/osism/testbed/.git/logs/refs/heads/\n<f..t...... src/github.com/osism/testbed/.git/logs/refs/heads/main\n.d..t...... src/github.com/osism/testbed/.git/logs/refs/remotes/\n.d..t...... src/github.com/osism/testbed/.git/logs/refs/remotes/origin/\n<f..t...... src/github.com/osism/testbed/.git/logs/refs/remotes/origin/HEAD\n.d..t...... src/github.com/osism/testbed/.git/objects/\n.d..t...... src/github.com/osism/testbed/.git/objects/00/\n.d..t...... src/github.com/osism/testbed/.git/objects/01/\n.d..t...... src/github.com/osism/testbed/.git/objects/02/\n.d..t...... src/github.com/osism/testbed/.git/objects/03/\n.d..t...... src/github.com/osism/testbed/.git/objects/04/\n.d..t...... src/github.com/osism/testbed/.git/objects/05/\n.d..t...... src/github.com/osism/testbed/.git/objects/06/\n.d..t...... src/github.com/osism/testbed/.git/objects/07/\n.d..t...... src/github.com/osism/testbed/.git/objects/08/\n.d..t...... src/github.com/osism/testbed/.git/objects/09/\n.d..t...... src/github.com/osism/testbed/.git/objects/0a/\n.d..t...... src/github.com/osism/testbed/.git/objects/0b/\n.d..t...... src/github.com/osism/testbed/.git/objects/0c/\n.d..t...... src/github.com/osism/testbed/.git/objects/0d/\n.d..t...... src/github.com/osism/testbed/.git/objects/0e/\n.d..t...... src/github.com/osism/testbed/.git/objects/0f/\n.d..t...... src/github.com/osism/testbed/.git/objects/10/\n.d..t...... src/github.com/osism/testbed/.git/objects/11/\n.d..t...... src/github.com/osism/testbed/.git/objects/12/\n.d..t...... src/github.com/osism/testbed/.git/objects/13/\n.d..t...... src/github.com/osism/testbed/.git/objects/14/\n.d..t...... src/github.com/osism/testbed/.git/objects/15/\n.d..t...... src/github.com/osism/testbed/.git/objects/16/\n.d..t...... src/github.com/osism/testbed/.git/objects/17/\n.d..t...... src/github.com/osism/testbed/.git/objects/18/\n.d..t...... src/github.com/osism/testbed/.git/objects/19/\n.d..t...... src/github.com/osism/testbed/.git/objects/1a/\n.d..t...... src/github.com/osism/testbed/.git/objects/1b/\n.d..t...... src/github.com/osism/testbed/.git/objects/1c/\n.d..t...... src/github.com/osism/testbed/.git/objects/1d/\n.d..t...... src/github.com/osism/testbed/.git/objects/1e/\n.d..t...... src/github.com/osism/testbed/.git/objects/1f/\n.d..t...... src/github.com/osism/testbed/.git/objects/20/\n.d..t...... src/github.com/osism/testbed/.git/objects/21/\n.d..t...... src/github.com/osism/testbed/.git/objects/22/\n.d..t...... src/github.com/osism/testbed/.git/objects/23/\n.d..t...... src/github.com/osism/testbed/.git/objects/24/\n.d..t...... src/github.com/osism/testbed/.git/objects/25/\n.d..t...... src/github.com/osism/testbed/.git/objects/26/\n.d..t...... src/github.com/osism/testbed/.git/objects/27/\n.d..t...... src/github.com/osism/testbed/.git/objects/28/\n.d..t...... src/github.com/osism/testbed/.git/objects/29/\n.d..t...... src/github.com/osism/testbed/.git/objects/2a/\n.d..t...... src/github.com/osism/testbed/.git/objects/2b/\n.d..t...... src/github.com/osism/testbed/.git/objects/2c/\n.d..t...... src/github.com/osism/testbed/.git/objects/2d/\n.d..t...... src/github.com/osism/testbed/.git/objects/2e/\n.d..t...... src/github.com/osism/testbed/.git/objects/2f/\n.d..t...... src/github.com/osism/testbed/.git/objects/30/\n.d..t...... src/github.com/osism/testbed/.git/objects/31/\n.d..t...... src/github.com/osism/testbed/.git/objects/32/\n.d..t...... src/github.com/osism/testbed/.git/objects/33/\n.d..t...... src/github.com/osism/testbed/.git/objects/34/\n.d..t...... src/github.com/osism/testbed/.git/objects/35/\n.d..t...... src/github.com/osism/testbed/.git/objects/36/\n.d..t...... src/github.com/osism/testbed/.git/objects/37/\n.d..t...... src/github.com/osism/testbed/.git/objects/38/\n.d..t...... src/github.com/osism/testbed/.git/objects/39/\n.d..t...... src/github.com/osism/testbed/.git/objects/3a/\n.d..t...... src/github.com/osism/testbed/.git/objects/3b/\n.d..t...... src/github.com/osism/testbed/.git/objects/3c/\n.d..t...... src/github.com/osism/testbed/.git/objects/3d/\n.d..t...... src/github.com/osism/testbed/.git/objects/3e/\n.d..t...... src/github.com/osism/testbed/.git/objects/3f/\n.d..t...... src/github.com/osism/testbed/.git/objects/40/\n.d..t...... src/github.com/osism/testbed/.git/objects/41/\n.d..t...... src/github.com/osism/testbed/.git/objects/42/\n.d..t...... src/github.com/osism/testbed/.git/objects/43/\n.d..t...... src/github.com/osism/testbed/.git/objects/44/\n.d..t...... src/github.com/osism/testbed/.git/objects/45/\n.d..t...... src/github.com/osism/testbed/.git/objects/46/\n.d..t...... src/github.com/osism/testbed/.git/objects/47/\n.d..t...... src/github.com/osism/testbed/.git/objects/48/\n.d..t...... src/github.com/osism/testbed/.git/objects/49/\n.d..t...... src/github.com/osism/testbed/.git/objects/4a/\n.d..t...... src/github.com/osism/testbed/.git/objects/4c/\n.d..t...... src/github.com/osism/testbed/.git/objects/4d/\n.d..t...... src/github.com/osism/testbed/.git/objects/4e/\n.d..t...... src/github.com/osism/testbed/.git/objects/4f/\n.d..t...... src/github.com/osism/testbed/.git/objects/50/\n.d..t...... src/github.com/osism/testbed/.git/objects/51/\n.d..t...... src/github.com/osism/testbed/.git/objects/52/\n.d..t...... src/github.com/osism/testbed/.git/objects/53/\n.d..t...... src/github.com/osism/testbed/.git/objects/54/\n.d..t...... src/github.com/osism/testbed/.git/objects/55/\n.d..t...... src/github.com/osism/testbed/.git/objects/56/\n.d..t...... src/github.com/osism/testbed/.git/objects/57/\n.d..t...... src/github.com/osism/testbed/.git/objects/58/\n.d..t...... src/github.com/osism/testbed/.git/objects/59/\n.d..t...... src/github.com/osism/testbed/.git/objects/5a/\n.d..t...... src/github.com/osism/testbed/.git/objects/5b/\n.d..t...... src/github.com/osism/testbed/.git/objects/5c/\n.d..t...... src/github.com/osism/testbed/.git/objects/5d/\n.d..t...... src/github.com/osism/testbed/.git/objects/5e/\n.d..t...... src/github.com/osism/testbed/.git/objects/5f/\n.d..t...... src/github.com/osism/testbed/.git/objects/60/\n.d..t...... src/github.com/osism/testbed/.git/objects/61/\n.d..t...... src/github.com/osism/testbed/.git/objects/64/\n.d..t...... src/github.com/osism/testbed/.git/objects/65/\n.d..t...... src/github.com/osism/testbed/.git/objects/66/\n.d..t...... src/github.com/osism/testbed/.git/objects/67/\n.d..t...... src/github.com/osism/testbed/.git/objects/68/\n.d..t...... src/github.com/osism/testbed/.git/objects/69/\n.d..t...... src/github.com/osism/testbed/.git/objects/6a/\n.d..t...... src/github.com/osism/testbed/.git/objects/6b/\n.d..t...... src/github.com/osism/testbed/.git/objects/6c/\n.d..t...... src/github.com/osism/testbed/.git/objects/6d/\n.d..t...... src/github.com/osism/testbed/.git/objects/6e/\n.d..t...... src/github.com/osism/testbed/.git/objects/6f/\n.d..t...... src/github.com/osism/testbed/.git/objects/70/\n.d..t...... src/github.com/osism/testbed/.git/objects/71/\n.d..t...... src/github.com/osism/testbed/.git/objects/72/\n.d..t...... src/github.com/osism/testbed/.git/objects/73/\n.d..t...... src/github.com/osism/testbed/.git/objects/74/\n.d..t...... src/github.com/osism/testbed/.git/objects/75/\n.d..t...... src/github.com/osism/testbed/.git/objects/76/\n.d..t...... src/github.com/osism/testbed/.git/objects/77/\n.d..t...... src/github.com/osism/testbed/.git/objects/78/\n.d..t...... src/github.com/osism/testbed/.git/objects/79/\n.d..t...... src/github.com/osism/testbed/.git/objects/7a/\n.d..t...... src/github.com/osism/testbed/.git/objects/7b/\n.d..t...... src/github.com/osism/testbed/.git/objects/7c/\n.d..t...... src/github.com/osism/testbed/.git/objects/7d/\n.d..t...... src/github.com/osism/testbed/.git/objects/7e/\n.d..t...... src/github.com/osism/testbed/.git/objects/7f/\n.d..t...... src/github.com/osism/testbed/.git/objects/80/\n.d..t...... src/github.com/osism/testbed/.git/objects/81/\n.d..t...... src/github.com/osism/testbed/.git/objects/82/\n.d..t...... src/github.com/osism/testbed/.git/objects/83/\n.d..t...... src/github.com/osism/testbed/.git/objects/84/\n.d..t...... src/github.com/osism/testbed/.git/objects/85/\n.d..t...... src/github.com/osism/testbed/.git/objects/86/\n.d..t...... src/github.com/osism/testbed/.git/objects/87/\n.d..t...... src/github.com/osism/testbed/.git/objects/8a/\n.d..t...... src/github.com/osism/testbed/.git/objects/8b/\n.d..t...... src/github.com/osism/testbed/.git/objects/8c/\n.d..t...... src/github.com/osism/testbed/.git/objects/8e/\n.d..t...... src/github.com/osism/testbed/.git/objects/8f/\n.d..t...... src/github.com/osism/testbed/.git/objects/90/\n.d..t...... src/github.com/osism/testbed/.git/objects/91/\n.d..t...... src/github.com/osism/testbed/.git/objects/92/\n.d..t...... src/github.com/osism/testbed/.git/objects/93/\n.d..t...... src/github.com/osism/testbed/.git/objects/94/\n.d..t...... src/github.com/osism/testbed/.git/objects/95/\n.d..t...... src/github.com/osism/testbed/.git/objects/96/\n.d..t...... src/github.com/osism/testbed/.git/objects/97/\n.d..t...... src/github.com/osism/testbed/.git/objects/98/\n.d..t...... src/github.com/osism/testbed/.git/objects/9a/\n.d..t...... src/github.com/osism/testbed/.git/objects/9c/\n.d..t...... src/github.com/osism/testbed/.git/objects/9e/\n.d..t...... src/github.com/osism/testbed/.git/objects/9f/\n.d..t...... src/github.com/osism/testbed/.git/objects/a0/\n.d..t...... src/github.com/osism/testbed/.git/objects/a1/\n.d..t...... src/github.com/osism/testbed/.git/objects/a2/\n.d..t...... src/github.com/osism/testbed/.git/objects/a3/\n.d..t...... src/github.com/osism/testbed/.git/objects/a4/\n.d..t...... src/github.com/osism/testbed/.git/objects/a5/\n.d..t...... src/github.com/osism/testbed/.git/objects/a6/\n.d..t...... src/github.com/osism/testbed/.git/objects/a7/\n.d..t...... src/github.com/osism/testbed/.git/objects/a8/\n.d..t...... src/github.com/osism/testbed/.git/objects/a9/\n.d..t...... src/github.com/osism/testbed/.git/objects/aa/\n.d..t...... src/github.com/osism/testbed/.git/objects/ab/\n.d..t...... src/github.com/osism/testbed/.git/objects/ac/\n.d..t...... src/github.com/osism/testbed/.git/objects/ad/\n.d..t...... src/github.com/osism/testbed/.git/objects/ae/\n.d..t...... src/github.com/osism/testbed/.git/objects/af/\n.d..t...... src/github.com/osism/testbed/.git/objects/b0/\n.d..t...... src/github.com/osism/testbed/.git/objects/b1/\n.d..t...... src/github.com/osism/testbed/.git/objects/b2/\n.d..t...... src/github.com/osism/testbed/.git/objects/b3/\n.d..t...... src/github.com/osism/testbed/.git/objects/b4/\n.d..t...... src/github.com/osism/testbed/.git/objects/b5/\n.d..t...... src/github.com/osism/testbed/.git/objects/b6/\n.d..t...... src/github.com/osism/testbed/.git/objects/b7/\n.d..t...... src/github.com/osism/testbed/.git/objects/b8/\n.d..t...... src/github.com/osism/testbed/.git/objects/b9/\n.d..t...... src/github.com/osism/testbed/.git/objects/ba/\n.d..t...... src/github.com/osism/testbed/.git/objects/bb/\n.d..t...... src/github.com/osism/testbed/.git/objects/bc/\n.d..t...... src/github.com/osism/testbed/.git/objects/bd/\n.d..t...... src/github.com/osism/testbed/.git/objects/be/\n.d..t...... src/github.com/osism/testbed/.git/objects/bf/\n.d..t...... src/github.com/osism/testbed/.git/objects/c0/\n.d..t...... src/github.com/osism/testbed/.git/objects/c1/\n.d..t...... src/github.com/osism/testbed/.git/objects/c2/\n.d..t...... src/github.com/osism/testbed/.git/objects/c3/\n.d..t...... src/github.com/osism/testbed/.git/objects/c4/\n.d..t...... src/github.com/osism/testbed/.git/objects/c5/\n.d..t...... src/github.com/osism/testbed/.git/objects/c6/\n.d..t...... src/github.com/osism/testbed/.git/objects/c7/\n.d..t...... src/github.com/osism/testbed/.git/objects/c8/\n.d..t...... src/github.com/osism/testbed/.git/objects/c9/\n.d..t...... src/github.com/osism/testbed/.git/objects/ca/\n.d..t...... src/github.com/osism/testbed/.git/objects/cb/\n.d..t...... src/github.com/osism/testbed/.git/objects/cc/\n.d..t...... src/github.com/osism/testbed/.git/objects/cd/\n.d..t...... src/github.com/osism/testbed/.git/objects/ce/\n.d..t...... src/github.com/osism/testbed/.git/objects/cf/\n.d..t...... src/github.com/osism/testbed/.git/objects/d0/\n.d..t...... src/github.com/osism/testbed/.git/objects/d2/\n.d..t...... src/github.com/osism/testbed/.git/objects/d3/\n.d..t...... src/github.com/osism/testbed/.git/objects/d5/\n.d..t...... src/github.com/osism/testbed/.git/objects/d6/\n.d..t...... src/github.com/osism/testbed/.git/objects/d7/\n.d..t...... src/github.com/osism/testbed/.git/objects/d8/\n.d..t...... src/github.com/osism/testbed/.git/objects/d9/\n.d..t...... src/github.com/osism/testbed/.git/objects/da/\n.d..t...... src/github.com/osism/testbed/.git/objects/db/\n.d..t...... src/github.com/osism/testbed/.git/objects/dc/\n.d..t...... src/github.com/osism/testbed/.git/objects/dd/\n.d..t...... src/github.com/osism/testbed/.git/objects/de/\n.d..t...... src/github.com/osism/testbed/.git/objects/df/\n.d..t...... src/github.com/osism/testbed/.git/objects/e0/\n.d..t...... src/github.com/osism/testbed/.git/objects/e2/\n.d..t...... src/github.com/osism/testbed/.git/objects/e3/\n.d..t...... src/github.com/osism/testbed/.git/objects/e4/\n.d..t...... src/github.com/osism/testbed/.git/objects/e5/\n.d..t...... src/github.com/osism/testbed/.git/objects/e6/\n.d..t...... src/github.com/osism/testbed/.git/objects/e7/\n.d..t...... src/github.com/osism/testbed/.git/objects/e8/\n.d..t...... src/github.com/osism/testbed/.git/objects/e9/\n.d..t...... src/github.com/osism/testbed/.git/objects/ea/\n.d..t...... src/github.com/osism/testbed/.git/objects/eb/\n.d..t...... src/github.com/osism/testbed/.git/objects/ec/\n.d..t...... src/github.com/osism/testbed/.git/objects/ed/\n.d..t...... src/github.com/osism/testbed/.git/objects/ee/\n.d..t...... src/github.com/osism/testbed/.git/objects/ef/\n.d..t...... src/github.com/osism/testbed/.git/objects/f0/\n.d..t...... src/github.com/osism/testbed/.git/objects/f1/\n.d..t...... src/github.com/osism/testbed/.git/objects/f2/\n.d..t...... src/github.com/osism/testbed/.git/objects/f3/\n.d..t...... src/github.com/osism/testbed/.git/objects/f4/\n.d..t...... src/github.com/osism/testbed/.git/objects/f5/\n.d..t...... src/github.com/osism/testbed/.git/objects/f6/\n.d..t...... src/github.com/osism/testbed/.git/objects/f8/\n.d..t...... src/github.com/osism/testbed/.git/objects/f9/\n.d..t...... src/github.com/osism/testbed/.git/objects/fa/\n.d..t...... src/github.com/osism/testbed/.git/objects/fb/\n.d..t...... src/github.com/osism/testbed/.git/objects/fc/\n.d..t...... src/github.com/osism/testbed/.git/objects/fd/\n.d..t...... src/github.com/osism/testbed/.git/objects/fe/\n.d..t...... src/github.com/osism/testbed/.git/objects/ff/\n.d..t...... src/github.com/osism/testbed/.git/objects/info/\n.d..t...... src/github.com/osism/testbed/.git/objects/pack/\n.d..t...... src/github.com/osism/testbed/.git/refs/\n.d..t...... src/github.com/osism/testbed/.github/\n<f..t...... src/github.com/osism/testbed/.github/renovate.json\n.d..t...... src/github.com/osism/testbed/.github/workflows/\n<f..t...... src/github.com/osism/testbed/.github/workflows/update-manager-images.yml\n.d..t...... src/github.com/osism/testbed/ansible/\n<f..t...... src/github.com/osism/testbed/ansible/check-local-versions.yml\n<f..t...... src/github.com/osism/testbed/ansible/localhost_inventory.yaml\n<f..t...... src/github.com/osism/testbed/ansible/manager-part-0.yml\n<f..t...... src/github.com/osism/testbed/ansible/manager-part-1.yml\n<f..t...... src/github.com/osism/testbed/ansible/manager-part-2.yml\n<f..t...... src/github.com/osism/testbed/ansible/manager-part-3.yml\n.d..t...... src/github.com/osism/testbed/containerlab/\n<f..t...... src/github.com/osism/testbed/containerlab/README.md\n<f..t...... src/github.com/osism/testbed/containerlab/osism.clab.yml\n.d..t...... src/github.com/osism/testbed/containerlab/files/\n<f..t...... src/github.com/osism/testbed/containerlab/files/daemons\n<f..t...... src/github.com/osism/testbed/containerlab/files/frr01.conf\n<f..t...... src/github.com/osism/testbed/containerlab/files/frr02.conf\n<f..t...... src/github.com/osism/testbed/containerlab/files/vtysh.conf\n.d..t...... src/github.com/osism/testbed/contrib/\n<f..t...... src/github.com/osism/testbed/contrib/analyze-tempest-logs.sh\n<f..t...... src/github.com/osism/testbed/contrib/analyze-zuul-logs.py\n<f..t...... src/github.com/osism/testbed/contrib/elasticsearch_random_data_load.sh\n<f..t...... src/github.com/osism/testbed/contrib/mysql_random_data_load.sh\n<f..t...... src/github.com/osism/testbed/contrib/mysql_random_data_load.sql\n<f..t...... src/github.com/osism/testbed/contrib/semver2.sh\n<f..t...... src/github.com/osism/testbed/contrib/setup-testbed.py\n.d..t...... src/github.com/osism/testbed/contrib/ansible/\n<f..t...... src/github.com/osism/testbed/contrib/ansible/clouds.yaml.sample\n<f..t...... src/github.com/osism/testbed/contrib/ansible/dns.yml\n<f..t...... src/github.com/osism/testbed/contrib/ansible/secure.yaml.sample\n.d..t...... src/github.com/osism/testbed/contrib/assets/\n<f..t...... src/github.com/osism/testbed/contrib/assets/operations-dashboard.png\n.d..t...... src/github.com/osism/testbed/contrib/integration-tests/\n<f..t...... src/github.com/osism/testbed/contrib/integration-tests/docker-compose.yaml\n<f..t...... src/github.com/osism/testbed/contrib/integration-tests/run_test.sh\n.d..t...... src/github.com/osism/testbed/contrib/integration-tests/tests/\n<f..t...... src/github.com/osism/testbed/contrib/integration-tests/tests/new_user.robot\n.d..t...... src/github.com/osism/testbed/contrib/nagios-plugins/\n<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_elasticsearch\n<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_galera_cluster\n<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_rabbitmq_cluster\n.d..t...... src/github.com/osism/testbed/contrib/openstack-cli/\n<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/README.md\n<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/clouds-public.yaml\n<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/clouds.yaml.example\n<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/secure.yaml.example\n.d..t...... src/github.com/osism/testbed/contrib/ownca/\n<f..t...... src/github.com/osism/testbed/contrib/ownca/README.md\n<f..t...... src/github.com/osism/testbed/contrib/ownca/create_ca.yml\n<f..t...... src/github.com/osism/testbed/contrib/ownca/create_manager.yml\n<f..t...... src/github.com/osism/testbed/contrib/ownca/create_wildcard.yml\n.d..t...... src/github.com/osism/testbed/environments/\n<f..t...... src/github.com/osism/testbed/environments/.vault_pass\n<f..t...... src/github.com/osism/testbed/environments/ansible.cfg\n<f..t...... src/github.com/osism/testbed/environments/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/images.yml\n<f..t...... src/github.com/osism/testbed/environments/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/ceph.test/\n<f..t...... src/github.com/osism/testbed/environments/ceph.test/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/ceph.test/images.yml\n<f..t...... src/github.com/osism/testbed/environments/ceph.test/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/ceph/\n<f..t...... src/github.com/osism/testbed/environments/ceph/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/ceph/images.yml\n<f..t...... src/github.com/osism/testbed/environments/ceph/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/custom/\n<f..t...... src/github.com/osism/testbed/environments/custom/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/images.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-bootstrap.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-netbox.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-pull-images.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-facts.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-migrate-proxysql.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-pull-images.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-wipe-partitions.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-workarounds-ceph.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-workarounds.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/custom/files/\n<f..t...... src/github.com/osism/testbed/environments/custom/files/baremetal-tenks-override.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_devices.fact\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_devices_all.fact\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_osd_devices.fact\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_osd_devices_all.fact\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_network_devices.fact\n.d..t...... src/github.com/osism/testbed/environments/custom/templates/\n<f..t...... src/github.com/osism/testbed/environments/custom/templates/baremetal-netbox-device.yml.j2\n<f..t...... src/github.com/osism/testbed/environments/custom/templates/ironic-interface.service.j2\n<f..t...... src/github.com/osism/testbed/environments/custom/templates/workarounds.service.j2\n<f..t...... src/github.com/osism/testbed/environments/custom/templates/workarounds.sh.j2\n.d..t...... src/github.com/osism/testbed/environments/generic/\n<f..t...... src/github.com/osism/testbed/environments/generic/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/generic/images.yml\n<f..t...... src/github.com/osism/testbed/environments/generic/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/infrastructure/\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/images.yml\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/infrastructure/files/\n.d..t...... src/github.com/osism/testbed/environments/infrastructure/files/ceph/\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/ceph/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/cert.crt\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/private_key.pem\n.d..t...... src/github.com/osism/testbed/environments/kolla/\n<f..t...... src/github.com/osism/testbed/environments/kolla/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/kolla/images.yml\n<f..t...... src/github.com/osism/testbed/environments/kolla/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/kolla/certificates/\n<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/haproxy-internal.pem\n<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/haproxy.pem\n.d..t...... src/github.com/osism/testbed/environments/kolla/certificates/ca/\n<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/ca/testbed.crt\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/galera.cnf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/global.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana.ini\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/keystone.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/event_pipeline.yaml\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/pipeline.yaml\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/ceph.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-scheduler.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-backup/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-backup/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/ceph.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/glance-api.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/gnocchi/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/gnocchi/ceph.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/provisioning.yaml\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/dashboards/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/dashboards/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/services.d/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-agent.initramfs\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-agent.kernel\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-conductor.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-dnsmasq.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/manila/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/manila/ceph.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron/ml2_conf.ini\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/ceph.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/nova-compute.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/client.cert-and-key.pem\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/client_ca.cert.pem\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/server_ca.cert.pem\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/server_ca.key.pem\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus.yml.d/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus.yml.d/50-ceph.yml\n.d..t...... src/github.com/osism/testbed/environments/kubernetes/\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/images.yml\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/playbook-nginx.yml\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/kubernetes/files/\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/files/nginx.yml\n.d..t...... src/github.com/osism/testbed/environments/manager/\n<f..t...... src/github.com/osism/testbed/environments/manager/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/manager/hosts\n<f..t...... src/github.com/osism/testbed/environments/manager/images.yml\n<f..t...... src/github.com/osism/testbed/environments/manager/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/manager/files/\n<f..t...... src/github.com/osism/testbed/environments/manager/files/conductor.yml\n.d..t...... src/github.com/osism/testbed/environments/monitoring/\n<f..t...... src/github.com/osism/testbed/environments/monitoring/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/monitoring/images.yml\n<f..t...... src/github.com/osism/testbed/environments/monitoring/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/openstack/\n<f..t...... src/github.com/osism/testbed/environments/openstack/clouds.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/id_rsa.test\n<f..t...... src/github.com/osism/testbed/environments/openstack/id_rsa.test.pub\n<f..t...... src/github.com/osism/testbed/environments/openstack/images.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/openrc\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-basic.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-ceph-rgw.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-flavors.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-images.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-test-identity.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-test.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/secrets.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/secure.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/testbed.pem\n.d..t...... src/github.com/osism/testbed/environments/openstack/files/\n.d..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/\n<f..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/exclude.lst\n<f..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/include-scs-compatible.lst\n.d..t...... src/github.com/osism/testbed/inventory/\n<f..t...... src/github.com/osism/testbed/inventory/10-custom\n<f..t...... src/github.com/osism/testbed/inventory/20-roles\n<f..t...... src/github.com/osism/testbed/inventory/99-overwrite\n.d..t...... src/github.com/osism/testbed/inventory/group_vars/\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-external-nodes.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-managers.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-nodes.yml\n.d..t...... src/github.com/osism/testbed/inventory/group_vars/all/\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/ansible.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/kolla.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/registries.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/vxlan.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-manager/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-manager/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-0/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-0/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-1/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-1/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-2/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-2/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-3/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-3/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-4/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-4/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-5/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-5/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-6/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-6/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-7/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-7/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-8/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-8/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-9/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-9/vars.yml\n.d..t...... src/github.com/osism/testbed/netbox/\n.d..t...... src/github.com/osism/testbed/netbox/devicetypes/\n.d..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/5835-54X-O-AC-F.yaml\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/7726-32X-O-AC-F.yaml\n.d..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/baremetal-device.yml\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/baremetal-housing.yml\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/manager.yml\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/node.yml\n.d..t...... src/github.com/osism/testbed/netbox/moduletypes/\n<f..t...... src/github.com/osism/testbed/netbox/moduletypes/.gitkeep\n.d..t...... src/github.com/osism/testbed/netbox/resources/\n<f..t...... src/github.com/osism/testbed/netbox/resources/100-initialise.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/200-rack-1000.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-manager.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-0.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-1.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-2.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-3.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-4.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-5.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-6.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-7.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-8.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-9.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-0.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-1.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-2.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-3.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-oob.yml\n.d..t...... src/github.com/osism/testbed/network/\n<f..t...... src/github.com/osism/testbed/network/iptables.sh\n<f..t...... src/github.com/osism/testbed/network/vxlan.sh\n.d..t...... src/github.com/osism/testbed/playbooks/\n<f..t...... src/github.com/osism/testbed/playbooks/cleanup.yml\n<f..t...... src/github.com/osism/testbed/playbooks/deploy.yml\n<f..t...... src/github.com/osism/testbed/playbooks/post.yml\n<f..t...... src/github.com/osism/testbed/playbooks/pre.yml\n<f..t...... src/github.com/osism/testbed/playbooks/update-stable.yml\n<f..t...... src/github.com/osism/testbed/playbooks/upgrade-stable.yml\n<f..t...... src/github.com/osism/testbed/playbooks/upgrade.yml\n.d..t...... src/github.com/osism/testbed/playbooks/templates/\n<f..t...... src/github.com/osism/testbed/playbooks/templates/local.env.j2\n.d..t...... src/github.com/osism/testbed/playbooks/vars/\n<f..t...... src/github.com/osism/testbed/playbooks/vars/mappings.yml\n<f..t...... src/github.com/osism/testbed/playbooks/vars/repositories.yml\n.d..t...... src/github.com/osism/testbed/scripts/\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap-services.sh\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap.sh\n<f..t...... src/github.com/osism/testbed/scripts/check-ceph.sh\n<f..t...... src/github.com/osism/testbed/scripts/check-services.sh\n<f..t...... src/github.com/osism/testbed/scripts/check.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-baremetal.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-ceph.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-in-a-nutshell.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-manager.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-services.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy.sh\n<f..t...... src/github.com/osism/testbed/scripts/disable-ara.sh\n<f..t...... src/github.com/osism/testbed/scripts/disable-local-registry.sh\n<f..t...... src/github.com/osism/testbed/scripts/enable-ara.sh\n<f..t...... src/github.com/osism/testbed/scripts/enable-resource-nodes.sh\n<f..t...... src/github.com/osism/testbed/scripts/include.sh\n<f..t...... src/github.com/osism/testbed/scripts/manager-version.sh\n<f..t...... src/github.com/osism/testbed/scripts/prepare-ceph-configuration.sh\n<f..t...... src/github.com/osism/testbed/scripts/prepare-wireguard-configuration.sh\n<f..t...... src/github.com/osism/testbed/scripts/pull-images.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-ceph-version.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-docker-registry.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-kolla-namespace.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-manager-version.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-openstack-version.sh\n<f..t...... src/github.com/osism/testbed/scripts/sync-configuration-repository.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade-ceph.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade-manager.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade-services.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade.sh\n<f..t...... src/github.com/osism/testbed/scripts/wrapper-gilt.sh\n.d..t...... src/github.com/osism/testbed/scripts/bootstrap/\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap/000-netbox.sh\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap/300-openstack.sh\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap/302-openstack-k8s-clusterapi-images.sh\n.d..t...... src/github.com/osism/testbed/scripts/check/\n<f..t...... src/github.com/osism/testbed/scripts/check/100-ceph-with-ansible.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/100-ceph-with-rook.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/200-infrastructure.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/300-openstack.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/301-openstack-domain-manager.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/301-openstack-scs-standards.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/302-openstack-scs-compatible.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/302-openstack-with-tempest.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/303-prometheus-alert-status.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/500-kubernetes.sh\n.d..t...... src/github.com/osism/testbed/scripts/customisations/\n<f..t...... src/github.com/osism/testbed/scripts/customisations/external-api.sh\n.d..t...... src/github.com/osism/testbed/scripts/deploy/\n<f..t...... src/github.com/osism/testbed/scripts/deploy/000-manager.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/001-helpers.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/100-ceph-with-ansible.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/100-ceph-with-rook.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/200-infrastructure.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/300-openstack.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/310-openstack-extended.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/320-openstack-minimal.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/400-monitoring.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/500-kubernetes.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/510-clusterapi.sh\n.d..t...... src/github.com/osism/testbed/scripts/upgrade/\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/100-ceph-with-ansible.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/100-ceph-with-rook.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/200-infrastructure.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/300-openstack.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/310-openstack-extended.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/320-openstack-minimal.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/400-monitoring.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/500-kubernetes.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/510-clusterapi.sh\n.d..t...... src/github.com/osism/testbed/terraform/\n<f..t...... src/github.com/osism/testbed/terraform/Makefile\n<f..t...... src/github.com/osism/testbed/terraform/README.md\n<f..t...... src/github.com/osism/testbed/terraform/backend.env.sample\n<f..t...... src/github.com/osism/testbed/terraform/backend.tf.sample\n<f..t...... src/github.com/osism/testbed/terraform/clouds-public.yaml\n<f..t...... src/github.com/osism/testbed/terraform/clouds.yaml.sample\n<f..t...... src/github.com/osism/testbed/terraform/local.env.sample\n<f..t...... src/github.com/osism/testbed/terraform/secure.yaml.sample\n<f..t...... src/github.com/osism/testbed/terraform/terraformrc.sample\n.d..t...... src/github.com/osism/testbed/terraform/environments/\n<f..t...... src/github.com/osism/testbed/terraform/environments/artcodix.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/cah-f1a.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ci-centos-stream-9.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ci-debian-12.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ci-ubuntu-24.04.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ci.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/cleura.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/cloudpod.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/default.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/exigo.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/gx-scs.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/huaweicloud.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/otc.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ovh.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/pluscloudopen.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/regiocloud.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/wavestack.tfvars\n.d..t...... src/github.com/osism/testbed/terraform/scripts/\n<f..t...... src/github.com/osism/testbed/terraform/scripts/cleanup.py\n*deleting   src/github.com/osism/testbed/ansible/ansible-manager-part-0.log\n*deleting   src/github.com/osism/testbed/ansible/.id_rsa.ci.pub\n*deleting   src/github.com/osism/testbed/ansible/.id_rsa.ci\n*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/ci/terraform.tfstate\n*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/ci/\n*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/\n*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_use_ephemeral_storage_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_boot_from_volume_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_boot_from_image_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_router_enable_snat_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_availability_zone_hints_router_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_availability_zone_hints_network_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/manager_boot_from_volume_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/manager_boot_from_image_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/\n*deleting   src/github.com/osism/testbed/terraform/customisations/neutron_floatingip_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/external_api_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/default_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/access_ipv6_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/access_ipv4_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/access_floatingip_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/terraform-provider-openstack_v3.4.0\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/README.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/LICENSE\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/CHANGELOG.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64.lock\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/terraform-provider-null\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/README.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/LICENSE\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/CHANGELOG.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64.lock\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/terraform-provider-local\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/README.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/LICENSE\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/CHANGELOG.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64.lock\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/environment\n*deleting   src/github.com/osism/testbed/terraform/.terraform/\n*deleting   src/github.com/osism/testbed/terraform/variables.tf\n*deleting   src/github.com/osism/testbed/terraform/provider.tf\n*deleting   src/github.com/osism/testbed/terraform/outputs.tf\n*deleting   src/github.com/osism/testbed/terraform/nova.tf\n*deleting   src/github.com/osism/testbed/terraform/nodes_boot_from_volume_override.tf\n*deleting   src/github.com/osism/testbed/terraform/nodes.tf\n*deleting   src/github.com/osism/testbed/terraform/neutron_floatingip_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/neutron_availability_zone_hints_router_override.tf\n*deleting   src/github.com/osism/testbed/terraform/neutron_availability_zone_hints_network_override.tf\n*deleting   src/github.com/osism/testbed/terraform/neutron.tf\n*deleting   src/github.com/osism/testbed/terraform/manager_boot_from_volume_override.tf\n*deleting   src/github.com/osism/testbed/terraform/manager.tf\n*deleting   src/github.com/osism/testbed/terraform/main.tf\n*deleting   src/github.com/osism/testbed/terraform/local.env\n*deleting   src/github.com/osism/testbed/terraform/inventory.ci\n*deleting   src/github.com/osism/testbed/terraform/default_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/data.tf\n*deleting   src/github.com/osism/testbed/terraform/ci.auto.tfvars\n*deleting   src/github.com/osism/testbed/terraform/access_floatingip_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/.terraform.lock.hcl\n*deleting   src/github.com/osism/testbed/terraform/.id_rsa.ci.pub\n*deleting   src/github.com/osism/testbed/terraform/.id_rsa.ci\n*deleting   src/github.com/osism/testbed/terraform/.deploy.ci\n*deleting   src/github.com/osism/testbed/terraform/.MANAGER_ADDRESS.ci\n",
                            "rc": 0,
                            "stdout_lines": [
                                ".d..t...... src/",
                                ".d..t...... src/github.com/",
                                ".d..t...... src/github.com/osism/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.ansible-lint",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.flake8",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.gitignore",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.yamllint.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.zuul.yaml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/CHANGELOG.md",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/LICENSE",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/README.md",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/galaxy.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/HEAD",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/ORIG_HEAD",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/config",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/config.worktree",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/description",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/index",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/packed-refs",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/branches/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/applypatch-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/commit-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/fsmonitor-watchman.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/post-update.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-applypatch.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-commit.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-merge-commit.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-push.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-rebase.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-receive.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/prepare-commit-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/push-to-checkout.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/update.sample",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/info/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/info/exclude",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/HEAD",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/heads/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/heads/main",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/origin/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/origin/HEAD",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/01/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/02/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/04/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/05/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/06/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/07/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/08/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/09/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/10/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/11/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/12/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/13/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/14/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/15/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/16/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/17/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/18/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/19/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/20/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/21/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/22/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/23/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/24/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/25/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/26/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/27/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/28/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/29/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/30/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/31/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/32/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/33/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/35/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/36/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/37/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/38/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/39/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/40/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/41/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/42/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/43/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/44/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/45/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/46/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/47/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/48/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/49/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/50/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/52/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/53/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/54/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/55/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/56/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/57/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/58/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/59/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/60/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/61/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/62/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/63/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/64/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/66/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/67/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/68/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/69/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/70/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/71/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/72/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/73/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/74/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/75/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/76/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/77/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/78/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/79/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/80/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/81/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/82/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/83/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/84/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/85/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/86/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/87/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/88/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/89/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/90/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/91/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/93/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/94/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/95/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/96/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/97/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/98/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a1/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a8/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/aa/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ab/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ac/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ad/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ae/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/af/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b1/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b8/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ba/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bb/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bc/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bd/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/be/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bf/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ca/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cb/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cc/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cd/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ce/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cf/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d1/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d8/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/da/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/db/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/dc/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/de/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/df/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e8/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ea/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/eb/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ec/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ed/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ee/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ef/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f1/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f8/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fa/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fb/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fc/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fd/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fe/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ff/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/info/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/pack/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/refs/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.github/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.github/publish-collection.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.github/renovate.json",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.github/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.github/templates/galaxy.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.github/workflows/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.github/workflows/publish-collection.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/meta/runtime.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/requirements.txt",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/collections.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/converge.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/molecule.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/certificates.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/cleanup.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/configfs.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/configuration.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/docker_compose.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/docker_login.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/facts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/firewall.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/hostname.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/ipmitool.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/k9s.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/kernel_modules.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/known_hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/limits.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/lynis.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/microcode.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/motd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/network.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/operator.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/packages.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/podman.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/proxy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/repository.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/resolvconf.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/runc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/services.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sosreport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sshconfig.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/state.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sysctl.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/systohc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/timezone.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/trivy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/user.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/certificates.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configfs.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/docker_compose.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/docker_login.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/facts.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/firewall.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hostname.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/ipmitool.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/k9s.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/kernel_modules.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/known_hosts.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/limits.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/microcode.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/operator.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/resolvconf.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/runc.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/services.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sosreport.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sshconfig.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/state.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sysctl.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/systohc.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/user.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/cloudinit.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/packages.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/packages_debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/services.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/timer_debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/timer_redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/git.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/type-block.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/type-template.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/interfaces.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/netplan.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/centos.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/debian.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/util.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/certificates.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/cleanup.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/configfs.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/configuration.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/docker_compose.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/docker_login.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/facts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/firewall.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/hostname.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/ipmitool.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/k9s.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/kernel_modules.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/known_hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/limits.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/lynis.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/microcode.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/motd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/network.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/operator.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/packages.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/podman.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/proxy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/repository.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/resolvconf.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/runc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/services.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sosreport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sshconfig.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/state.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sysctl.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/systohc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/timezone.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/trivy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/user.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/certificates.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/cleanup.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/configfs.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/configuration.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/docker_compose.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/facts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/firewall.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/hostname.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/ipmitool.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/k9s.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/kernel_modules.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/lynis.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/microcode.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/motd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/operator.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/packages.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/podman.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/proxy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/repository.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/resolvconf.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/runc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/services.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sosreport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sshconfig.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/state.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sysctl.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/systohc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/timezone.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/trivy.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/plugins/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/plugins/callback/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/plugins/callback/still_alive.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/templates/certificates.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/cloudinit.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/packages-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/packages-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/services-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/services-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/timers-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/timers-RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/git.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/update.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config-git-netbox-private-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config-git-private-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/files/docker-compose",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/docker-compose.service.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/osism.target.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/templates/ufw.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/templates/config.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-block.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-fast.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-local.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-template.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/hosts-block.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/hosts-fast.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/hosts-template.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/templates/module-load.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/write-scanned.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/write-static.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/limits.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/configure-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/configure-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/templates/motd.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-interfaces.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-netplan.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-networkd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/dummy-interfaces.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/interfaces-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/interfaces-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/netplan-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/netplan-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/network-extra-init.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/vxlan-interfaces.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/dummy.netdev.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/network-extra-init.service.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/network-extra-init.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/vxlan.netdev.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/vxlan.network.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/device.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/interfaces.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/01-osism.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/interfaces.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/interfaces.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/templates/operator-sudoers.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/package-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/package-RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/deploy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/templates/proxy-apt.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/",
                                ".L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/AlmaLinux.yml -> RedHat.yml",
                                ".L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/CentOS.yml -> RedHat.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Debian.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/RedHat.yml",
                                ".L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Rocky.yml -> RedHat.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Ubuntu.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/99osism.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/source.list.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/ubuntu.sources.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/AlmaLinux-amd64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/CentOS-x86_64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Debian-x86_64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Rocky-x86_64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Ubuntu-aarch64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Ubuntu-x86_64.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-resolv.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/templates/resolved.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/services/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/services/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/services/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/services/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/CentOS-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/Debian-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/Ubuntu-dist.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/templates/config.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/state/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/state/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/state/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/state/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/sysctl.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/remote-key.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/type-default.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/type-keyfile.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/templates/config.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.ansible-lint",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.flake8",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.gitignore",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.yamllint.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.zuul.yaml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/CHANGELOG.md",
                                "<f..t...... src/github.com/osism/ansible-collection-services/LICENSE",
                                "<f..t...... src/github.com/osism/ansible-collection-services/README.md",
                                "<f..t...... src/github.com/osism/ansible-collection-services/galaxy.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/HEAD",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/ORIG_HEAD",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/config",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/config.worktree",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/description",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/index",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/packed-refs",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/branches/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/hooks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/applypatch-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/commit-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/fsmonitor-watchman.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/post-update.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-applypatch.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-commit.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-merge-commit.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-push.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-rebase.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-receive.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/prepare-commit-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/push-to-checkout.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/update.sample",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/info/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/info/exclude",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/logs/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/HEAD",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/heads/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/heads/main",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/origin/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/origin/HEAD",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/00/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/01/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/02/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/03/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/05/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/07/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/08/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/09/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/10/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/11/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/12/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/13/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/14/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/16/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/19/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1d/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1e/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1f/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/20/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/21/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/22/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/23/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/24/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/25/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/26/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/27/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/28/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/29/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2d/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2e/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/30/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/31/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/32/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/33/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/34/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/38/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3e/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3f/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/40/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/41/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/42/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/43/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/44/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/45/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/49/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4e/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4f/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/50/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/51/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/53/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/54/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/55/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/56/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/57/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/58/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/59/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/5a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/5b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/60/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/61/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/63/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/64/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/65/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/66/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/67/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/68/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/69/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6f/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/70/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/71/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/72/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/73/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/74/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/75/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/76/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/77/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/79/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7d/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7e/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7f/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/80/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/81/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/82/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/83/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/84/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/85/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/87/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/88/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/89/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8d/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8e/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/91/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/92/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/93/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/94/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/95/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/96/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/97/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/98/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9f/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a0/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a1/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a2/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a3/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a4/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a5/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a8/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/aa/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ac/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ad/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ae/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/af/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b0/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b1/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b2/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b3/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b4/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b6/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b7/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b8/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b9/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ba/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/bb/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/bc/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/bd/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c0/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c2/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c3/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c4/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c5/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c6/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c9/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/cb/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/cc/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ce/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/cf/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d0/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d1/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d2/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d4/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d5/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d7/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d8/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d9/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/da/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/db/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/dc/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/dd/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/de/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/df/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e0/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e1/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e3/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e5/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e6/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e7/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e8/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e9/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ea/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/eb/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ec/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ed/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ee/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ef/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f0/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f2/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f3/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f6/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f7/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f8/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fa/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fd/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fe/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ff/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/info/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/pack/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/refs/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.github/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.github/publish-collection.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.github/renovate.json",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.github/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.github/templates/galaxy.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.github/workflows/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.github/workflows/publish-collection.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/meta/runtime.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/requirements.txt",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/collections.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/converge.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/molecule.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/adminer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/auditd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/cephclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/cgit.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/chrony.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/clamav.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/containerd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/dnsdist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/dnsmasq.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/docker.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/fail2ban.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/falco.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/frr.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/gnmic.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/hddtemp.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/homer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/httpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/journald.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/lldpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/manager.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netbird.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netdata.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/nexus.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/openstackclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/opentelemetry_collector.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/osquery.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/phpmyadmin.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/rng.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/rsyslog.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/scaphandre.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/smartd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/squid.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/sshd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/stepca.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/substation.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/teleport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/thanos_sidecar.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/traefik.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/tuned.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/wazuh_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/wireguard.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/zabbix_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/zuul.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/adminer.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/auditd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cgit.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/dnsdist.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/dnsmasq.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/fail2ban.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/frr.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/gnmic.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/homer.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/httpd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/journald.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/lldpd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/manager.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbox.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/nexus.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/opentelemetry_collector.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/phpmyadmin.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/rng.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/rsyslog.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/scaphandre.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/smartd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/squid.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/sshd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/stepca.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/substation.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/thanos_sidecar.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/traefik.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/tuned.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wireguard.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zuul.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/container.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/package.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/package_debian.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/debian.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/container.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/package-debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/package-redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/util.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/adminer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/auditd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/cephclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/cgit.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/chrony.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/clamav.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/containerd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/dnsdist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/dnsmasq.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/docker.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/fail2ban.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/falco.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/frr.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/gnmic.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/hddtemp.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/homer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/httpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/journald.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/lldpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/manager.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netbird.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netdata.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/nexus.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/openstackclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/opentelemetry_collector.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/osquery.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/phpmyadmin.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/rng.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/rsyslog.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/scaphandre.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/smartd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/squid.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/sshd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/stepca.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/substation.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/teleport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/thanos_sidecar.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/traefik.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/tuned.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/wazuh_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/wireguard.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/zabbix_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/zuul.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/adminer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/auditd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/cephclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/cgit.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/chrony.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/clamav.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/containerd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/dnsdist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/dnsmasq.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/docker.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/fail2ban.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/falco.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/frr.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/gnmic.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/hddtemp.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/homer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/httpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/journald.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/lldpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/netbird.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/netdata.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/osquery.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/rng.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/rsyslog.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/smartd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/sshd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/teleport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/tuned.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/wazuh_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/wireguard.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/zabbix_agent.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/plugins/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/plugins/filter/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/plugins/filter/address.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/plugins/modules/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_container_facts.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_docker.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_toolbox.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/auditd.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/rules/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/rules/20-neo23x0.rules.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-container.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-package-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-package-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/container.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/package-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/package-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/rook.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/ceph.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/keyring.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/ceph-authtool.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/ceph.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/crushtool.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rados.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/radosgw-admin.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rbd.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rook.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/cgitrc.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/httpd-cgit.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/templates/chrony.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/initialize.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/templates/config.toml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/CentOS-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/Debian-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/Ubuntu-dist.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/dnsdist.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/dnsmasq.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/.gitattributes",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/limits.conf",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/docker_containers.fact",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/docker_images.fact",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/bootstrap.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/facts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/install-docker-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/install-docker-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/storage-block.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/storage-zram.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/config.json.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/containerd_config.toml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/daemon.json.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/overlay.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/templates/sshd.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/.gitignore",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/files/falco.service",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-RedHat-family_broken.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/daemons.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_k3s_cilium.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_leaf.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_loadbalancer.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_loadbalancer_external_uplink.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_test.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_yrzn001.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_yrzn_metalbox.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_yrzn_network.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/vtysh.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/gnmic.yaml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/RedHat-family/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/RedHat-family/hddtemp",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/service-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/service-RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/config.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/files/prepare-ironic-volume.sh",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/ironic.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/sonic-ztp.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/htaccess.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/httpd.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/post_install.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/ztp.json.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/templates/journald.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/config.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-ansible.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-ara.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-celery.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-frontend.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-scripts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-vault.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-wrapper.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/initialize-vault.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/initialize.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/verify-versions.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/check-service-manager-versions.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-netbox-secret.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-ssh-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-vault-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/manager.service.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/vault.hcl.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/all.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ansible.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ara-server.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ara.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/client.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/conductor.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/frontend.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/inventory-reconciler.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/listener.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/mariadb.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/netbox.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/openstack.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/postgres.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/vault.env.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/scripts/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/scripts/osism-include.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/cilium.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/flux.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/hubble.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-ansible.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-ceph.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-custom.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-generic.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-infrastructure.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-kolla.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-manager.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-monitoring.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-openstack.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-run-without-secrets.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-run.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-state.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-update-docker.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-update-manager.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-validate.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/scripts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/wrapper.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/startup-scripts/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/startup-scripts/270_tags.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config-netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config-postgres.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/rescue-service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/restart-service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/wait-for-healthy-service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-netbox-database.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-netbox-secret.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-postgres-secret.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/configuration.py.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/netbox.service.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/nginx-unit-1.26.json.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/nginx-unit.json.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/postgresql.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/netbox.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/postgres.env.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/custom_fields.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/device_roles.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/device_types.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/groups.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/manufacturers.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/object_permissions.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/prefix_vlan_roles.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/sites.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/tags.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/users.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/vars/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/client.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/server.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/netdata.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/stream.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/create_repos_from_list.groovy",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/setup_http_proxy.groovy",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/setup_realms.groovy",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/update_admin_password.groovy",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/call-script.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/declare-script.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/initialize.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.properties.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.service.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/container-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/container-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/package-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/package-RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/openstack.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/otel-collector-config.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rng/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rng/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rng/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/additional-log-server.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/fluentd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/logrotate.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/70-fluentd.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/71-additional-log-server.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/rsyslog-logrotate.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/rsyslog.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/CentOS-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/Debian-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/Ubuntu-dist.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/files/smartmontools",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/osism.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/config-wrapper.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/provisioner-acme.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/stepca.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/stepca.service.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/wrapper/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/wrapper/step.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/files/entrypoint.sh",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/substation.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/templates/teleport.yaml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/certificates.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/dynamic.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik-cert.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/client.keys.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/ossec.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/client.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/server.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/templates/zabbix_agent2.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/builder-logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/executor-logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/launcher-logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/mime.types",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/openssl.cnf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/web-logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/zk-ca.sh",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/copy_tasks.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/generate_certs_and_keys.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/template_tasks.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/docker-compose.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/httpd.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/main.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/nodepool.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/wait-to-start-gearman.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/wait-to-start.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/zoo.cfg.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/zuul.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/executor.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/mariadb.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/nodepool_builder.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/nodepool_launcher.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/scheduler.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/web.env.j2",
                                ".d..t...... src/github.com/osism/terraform-base/",
                                "<f..t...... src/github.com/osism/terraform-base/.yamllint.yml",
                                "<f..t...... src/github.com/osism/terraform-base/.zuul.yaml",
                                "<f..t...... src/github.com/osism/terraform-base/LICENSE",
                                "<f..t...... src/github.com/osism/terraform-base/README.md",
                                ".d..t...... src/github.com/osism/terraform-base/.git/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/HEAD",
                                "<f..t...... src/github.com/osism/terraform-base/.git/ORIG_HEAD",
                                "<f..t...... src/github.com/osism/terraform-base/.git/config",
                                "<f..t...... src/github.com/osism/terraform-base/.git/config.worktree",
                                "<f..t...... src/github.com/osism/terraform-base/.git/description",
                                "<f..t...... src/github.com/osism/terraform-base/.git/index",
                                "<f..t...... src/github.com/osism/terraform-base/.git/packed-refs",
                                ".d..t...... src/github.com/osism/terraform-base/.git/branches/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/hooks/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/applypatch-msg.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/commit-msg.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/fsmonitor-watchman.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/post-update.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-applypatch.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-commit.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-merge-commit.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-push.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-rebase.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-receive.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/prepare-commit-msg.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/push-to-checkout.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/update.sample",
                                ".d..t...... src/github.com/osism/terraform-base/.git/info/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/info/exclude",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/logs/HEAD",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/refs/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/refs/heads/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/logs/refs/heads/main",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/origin/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/origin/HEAD",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/00/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/01/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/02/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/05/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/06/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/07/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/08/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/0a/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/0b/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/0d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/0f/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/10/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/12/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/15/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/16/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/17/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/1b/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/1c/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/1f/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/20/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/21/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/24/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/26/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/29/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/2c/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/36/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/37/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/3a/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/3d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/41/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/43/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/46/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/47/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/48/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/49/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/4b/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/4d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/52/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/55/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/56/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/58/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/5d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/61/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/63/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/66/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/67/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/68/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/6d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/70/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/72/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/76/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/7c/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/7f/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/80/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/81/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/82/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/83/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/84/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/85/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/86/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/88/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/8c/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/8d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/90/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/91/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/92/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/95/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/96/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/98/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/9b/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/9e/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a2/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a3/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a6/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a7/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a9/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/aa/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/af/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b0/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b3/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b4/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b8/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b9/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ba/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/bc/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/bf/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c0/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c1/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c4/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c7/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c8/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/cb/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ce/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/cf/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/d2/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/d3/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/d6/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/d7/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/d9/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/da/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/db/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/dc/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/dd/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/de/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/df/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/e1/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/e2/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/e3/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/e4/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/e8/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ea/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ee/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ef/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/f1/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/f2/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/f4/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/f9/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/fb/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ff/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/info/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/pack/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/refs/",
                                ".d..t...... src/github.com/osism/terraform-base/.github/",
                                "<f..t...... src/github.com/osism/terraform-base/.github/renovate.json",
                                ".d..t...... src/github.com/osism/terraform-base/playbooks/",
                                "<f..t...... src/github.com/osism/terraform-base/playbooks/validate.yml",
                                ".d..t...... src/github.com/osism/terraform-base/testbed-default/",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/data.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/main.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/manager.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/neutron.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/nodes.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/nova.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/outputs.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/provider.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/variables.tf",
                                ".d..t...... src/github.com/osism/terraform-base/testbed-default/customisations/",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_floatingip_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_ipv4_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_ipv6_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/default_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/external_api_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/neutron_floatingip_custom.tf",
                                ".d..t...... src/github.com/osism/terraform-base/testbed-default/environments/",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/environments/default.tfvars",
                                ".d..t...... src/github.com/osism/terraform-base/testbed-default/overrides/",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/manager_boot_from_image_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/manager_boot_from_volume_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_availability_zone_hints_network_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_availability_zone_hints_router_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_router_enable_snat_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_boot_from_image_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_boot_from_volume_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_use_ephemeral_storage_override.tf",
                                ".d..t...... src/github.com/osism/testbed/",
                                "<f..t...... src/github.com/osism/testbed/.ansible-lint",
                                "<f..t...... src/github.com/osism/testbed/.flake8",
                                "<f..t...... src/github.com/osism/testbed/.gitignore",
                                "<f..t...... src/github.com/osism/testbed/.yamllint.yml",
                                "<f..t...... src/github.com/osism/testbed/.zuul.yaml",
                                "<f..t...... src/github.com/osism/testbed/LICENSE",
                                "<f..t...... src/github.com/osism/testbed/Makefile",
                                "<f..t...... src/github.com/osism/testbed/README.md",
                                "<f..t...... src/github.com/osism/testbed/ansible.cfg",
                                "<f..t...... src/github.com/osism/testbed/gilt.yml",
                                "<f..t...... src/github.com/osism/testbed/requirements.txt",
                                "<f..t...... src/github.com/osism/testbed/requirements.yml",
                                "<f..t...... src/github.com/osism/testbed/tox.ini",
                                ".d..t...... src/github.com/osism/testbed/.git/",
                                "<f..t...... src/github.com/osism/testbed/.git/HEAD",
                                "<f..t...... src/github.com/osism/testbed/.git/ORIG_HEAD",
                                "<f..t...... src/github.com/osism/testbed/.git/config",
                                "<f..t...... src/github.com/osism/testbed/.git/config.worktree",
                                "<f..t...... src/github.com/osism/testbed/.git/description",
                                "<f..t...... src/github.com/osism/testbed/.git/index",
                                "<f..t...... src/github.com/osism/testbed/.git/packed-refs",
                                ".d..t...... src/github.com/osism/testbed/.git/branches/",
                                ".d..t...... src/github.com/osism/testbed/.git/hooks/",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/applypatch-msg.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/commit-msg.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/fsmonitor-watchman.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/post-update.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-applypatch.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-commit.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-merge-commit.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-push.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-rebase.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-receive.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/prepare-commit-msg.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/push-to-checkout.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/update.sample",
                                ".d..t...... src/github.com/osism/testbed/.git/info/",
                                "<f..t...... src/github.com/osism/testbed/.git/info/exclude",
                                ".d..t...... src/github.com/osism/testbed/.git/logs/",
                                "<f..t...... src/github.com/osism/testbed/.git/logs/HEAD",
                                ".d..t...... src/github.com/osism/testbed/.git/logs/refs/",
                                ".d..t...... src/github.com/osism/testbed/.git/logs/refs/heads/",
                                "<f..t...... src/github.com/osism/testbed/.git/logs/refs/heads/main",
                                ".d..t...... src/github.com/osism/testbed/.git/logs/refs/remotes/",
                                ".d..t...... src/github.com/osism/testbed/.git/logs/refs/remotes/origin/",
                                "<f..t...... src/github.com/osism/testbed/.git/logs/refs/remotes/origin/HEAD",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/00/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/01/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/02/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/03/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/04/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/05/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/06/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/07/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/08/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/09/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/10/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/11/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/12/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/13/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/14/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/15/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/16/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/17/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/18/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/19/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/20/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/21/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/22/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/23/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/24/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/25/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/26/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/27/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/28/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/29/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/30/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/31/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/32/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/33/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/34/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/35/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/36/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/37/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/38/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/39/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/40/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/41/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/42/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/43/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/44/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/45/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/46/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/47/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/48/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/49/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/50/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/51/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/52/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/53/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/54/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/55/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/56/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/57/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/58/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/59/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/60/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/61/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/64/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/65/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/66/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/67/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/68/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/69/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/70/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/71/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/72/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/73/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/74/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/75/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/76/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/77/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/78/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/79/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/80/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/81/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/82/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/83/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/84/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/85/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/86/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/87/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/90/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/91/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/92/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/93/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/94/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/95/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/96/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/97/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/98/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/9a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/9c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/9e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/9f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a1/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a4/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a7/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/aa/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ab/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ac/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ad/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ae/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/af/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b1/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b4/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b7/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ba/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/bb/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/bc/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/bd/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/be/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/bf/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c1/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c4/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c7/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ca/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/cb/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/cc/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/cd/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ce/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/cf/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d7/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/da/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/db/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/dc/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/dd/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/de/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/df/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e4/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e7/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ea/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/eb/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ec/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ed/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ee/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ef/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f1/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f4/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fa/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fb/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fc/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fd/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fe/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ff/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/info/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/pack/",
                                ".d..t...... src/github.com/osism/testbed/.git/refs/",
                                ".d..t...... src/github.com/osism/testbed/.github/",
                                "<f..t...... src/github.com/osism/testbed/.github/renovate.json",
                                ".d..t...... src/github.com/osism/testbed/.github/workflows/",
                                "<f..t...... src/github.com/osism/testbed/.github/workflows/update-manager-images.yml",
                                ".d..t...... src/github.com/osism/testbed/ansible/",
                                "<f..t...... src/github.com/osism/testbed/ansible/check-local-versions.yml",
                                "<f..t...... src/github.com/osism/testbed/ansible/localhost_inventory.yaml",
                                "<f..t...... src/github.com/osism/testbed/ansible/manager-part-0.yml",
                                "<f..t...... src/github.com/osism/testbed/ansible/manager-part-1.yml",
                                "<f..t...... src/github.com/osism/testbed/ansible/manager-part-2.yml",
                                "<f..t...... src/github.com/osism/testbed/ansible/manager-part-3.yml",
                                ".d..t...... src/github.com/osism/testbed/containerlab/",
                                "<f..t...... src/github.com/osism/testbed/containerlab/README.md",
                                "<f..t...... src/github.com/osism/testbed/containerlab/osism.clab.yml",
                                ".d..t...... src/github.com/osism/testbed/containerlab/files/",
                                "<f..t...... src/github.com/osism/testbed/containerlab/files/daemons",
                                "<f..t...... src/github.com/osism/testbed/containerlab/files/frr01.conf",
                                "<f..t...... src/github.com/osism/testbed/containerlab/files/frr02.conf",
                                "<f..t...... src/github.com/osism/testbed/containerlab/files/vtysh.conf",
                                ".d..t...... src/github.com/osism/testbed/contrib/",
                                "<f..t...... src/github.com/osism/testbed/contrib/analyze-tempest-logs.sh",
                                "<f..t...... src/github.com/osism/testbed/contrib/analyze-zuul-logs.py",
                                "<f..t...... src/github.com/osism/testbed/contrib/elasticsearch_random_data_load.sh",
                                "<f..t...... src/github.com/osism/testbed/contrib/mysql_random_data_load.sh",
                                "<f..t...... src/github.com/osism/testbed/contrib/mysql_random_data_load.sql",
                                "<f..t...... src/github.com/osism/testbed/contrib/semver2.sh",
                                "<f..t...... src/github.com/osism/testbed/contrib/setup-testbed.py",
                                ".d..t...... src/github.com/osism/testbed/contrib/ansible/",
                                "<f..t...... src/github.com/osism/testbed/contrib/ansible/clouds.yaml.sample",
                                "<f..t...... src/github.com/osism/testbed/contrib/ansible/dns.yml",
                                "<f..t...... src/github.com/osism/testbed/contrib/ansible/secure.yaml.sample",
                                ".d..t...... src/github.com/osism/testbed/contrib/assets/",
                                "<f..t...... src/github.com/osism/testbed/contrib/assets/operations-dashboard.png",
                                ".d..t...... src/github.com/osism/testbed/contrib/integration-tests/",
                                "<f..t...... src/github.com/osism/testbed/contrib/integration-tests/docker-compose.yaml",
                                "<f..t...... src/github.com/osism/testbed/contrib/integration-tests/run_test.sh",
                                ".d..t...... src/github.com/osism/testbed/contrib/integration-tests/tests/",
                                "<f..t...... src/github.com/osism/testbed/contrib/integration-tests/tests/new_user.robot",
                                ".d..t...... src/github.com/osism/testbed/contrib/nagios-plugins/",
                                "<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_elasticsearch",
                                "<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_galera_cluster",
                                "<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_rabbitmq_cluster",
                                ".d..t...... src/github.com/osism/testbed/contrib/openstack-cli/",
                                "<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/README.md",
                                "<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/clouds-public.yaml",
                                "<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/clouds.yaml.example",
                                "<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/secure.yaml.example",
                                ".d..t...... src/github.com/osism/testbed/contrib/ownca/",
                                "<f..t...... src/github.com/osism/testbed/contrib/ownca/README.md",
                                "<f..t...... src/github.com/osism/testbed/contrib/ownca/create_ca.yml",
                                "<f..t...... src/github.com/osism/testbed/contrib/ownca/create_manager.yml",
                                "<f..t...... src/github.com/osism/testbed/contrib/ownca/create_wildcard.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/",
                                "<f..t...... src/github.com/osism/testbed/environments/.vault_pass",
                                "<f..t...... src/github.com/osism/testbed/environments/ansible.cfg",
                                "<f..t...... src/github.com/osism/testbed/environments/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/ceph.test/",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph.test/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph.test/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph.test/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/ceph/",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/custom/",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-bootstrap.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-netbox.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-pull-images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-facts.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-migrate-proxysql.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-pull-images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-wipe-partitions.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-workarounds-ceph.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-workarounds.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/custom/files/",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/baremetal-tenks-override.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_devices.fact",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_devices_all.fact",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_osd_devices.fact",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_osd_devices_all.fact",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_network_devices.fact",
                                ".d..t...... src/github.com/osism/testbed/environments/custom/templates/",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/templates/baremetal-netbox-device.yml.j2",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/templates/ironic-interface.service.j2",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/templates/workarounds.service.j2",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/templates/workarounds.sh.j2",
                                ".d..t...... src/github.com/osism/testbed/environments/generic/",
                                "<f..t...... src/github.com/osism/testbed/environments/generic/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/generic/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/generic/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/infrastructure/",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/infrastructure/files/",
                                ".d..t...... src/github.com/osism/testbed/environments/infrastructure/files/ceph/",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/ceph/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/cert.crt",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/private_key.pem",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/certificates/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/haproxy-internal.pem",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/haproxy.pem",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/certificates/ca/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/ca/testbed.crt",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/galera.cnf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/global.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana.ini",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/keystone.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/event_pipeline.yaml",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/pipeline.yaml",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/ceph.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-scheduler.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-backup/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-backup/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/ceph.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/glance-api.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/gnocchi/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/gnocchi/ceph.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/provisioning.yaml",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/dashboards/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/dashboards/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/services.d/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-agent.initramfs",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-agent.kernel",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-conductor.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-dnsmasq.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/manila/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/manila/ceph.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron/ml2_conf.ini",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/ceph.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/nova-compute.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/client.cert-and-key.pem",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/client_ca.cert.pem",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/server_ca.cert.pem",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/server_ca.key.pem",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus.yml.d/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus.yml.d/50-ceph.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/kubernetes/",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/playbook-nginx.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/kubernetes/files/",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/files/nginx.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/manager/",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/hosts",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/manager/files/",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/files/conductor.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/monitoring/",
                                "<f..t...... src/github.com/osism/testbed/environments/monitoring/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/monitoring/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/monitoring/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/openstack/",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/clouds.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/id_rsa.test",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/id_rsa.test.pub",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/openrc",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-basic.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-ceph-rgw.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-flavors.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-test-identity.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-test.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/secrets.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/secure.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/testbed.pem",
                                ".d..t...... src/github.com/osism/testbed/environments/openstack/files/",
                                ".d..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/exclude.lst",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/include-scs-compatible.lst",
                                ".d..t...... src/github.com/osism/testbed/inventory/",
                                "<f..t...... src/github.com/osism/testbed/inventory/10-custom",
                                "<f..t...... src/github.com/osism/testbed/inventory/20-roles",
                                "<f..t...... src/github.com/osism/testbed/inventory/99-overwrite",
                                ".d..t...... src/github.com/osism/testbed/inventory/group_vars/",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-external-nodes.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-managers.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-nodes.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/group_vars/all/",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/ansible.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/kolla.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/registries.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/vxlan.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-manager/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-manager/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-0/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-0/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-1/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-1/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-2/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-2/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-3/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-3/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-4/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-4/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-5/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-5/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-6/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-6/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-7/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-7/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-8/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-8/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-9/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-9/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/netbox/",
                                ".d..t...... src/github.com/osism/testbed/netbox/devicetypes/",
                                ".d..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/5835-54X-O-AC-F.yaml",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/7726-32X-O-AC-F.yaml",
                                ".d..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/baremetal-device.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/baremetal-housing.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/manager.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/node.yml",
                                ".d..t...... src/github.com/osism/testbed/netbox/moduletypes/",
                                "<f..t...... src/github.com/osism/testbed/netbox/moduletypes/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/netbox/resources/",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/100-initialise.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/200-rack-1000.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-manager.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-0.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-1.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-2.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-3.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-4.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-5.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-6.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-7.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-8.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-9.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-0.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-1.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-2.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-3.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-oob.yml",
                                ".d..t...... src/github.com/osism/testbed/network/",
                                "<f..t...... src/github.com/osism/testbed/network/iptables.sh",
                                "<f..t...... src/github.com/osism/testbed/network/vxlan.sh",
                                ".d..t...... src/github.com/osism/testbed/playbooks/",
                                "<f..t...... src/github.com/osism/testbed/playbooks/cleanup.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/deploy.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/post.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/pre.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/update-stable.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/upgrade-stable.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/upgrade.yml",
                                ".d..t...... src/github.com/osism/testbed/playbooks/templates/",
                                "<f..t...... src/github.com/osism/testbed/playbooks/templates/local.env.j2",
                                ".d..t...... src/github.com/osism/testbed/playbooks/vars/",
                                "<f..t...... src/github.com/osism/testbed/playbooks/vars/mappings.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/vars/repositories.yml",
                                ".d..t...... src/github.com/osism/testbed/scripts/",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap-services.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check-ceph.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check-services.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-baremetal.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-ceph.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-in-a-nutshell.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-manager.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-services.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/disable-ara.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/disable-local-registry.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/enable-ara.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/enable-resource-nodes.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/include.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/manager-version.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/prepare-ceph-configuration.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/prepare-wireguard-configuration.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/pull-images.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-ceph-version.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-docker-registry.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-kolla-namespace.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-manager-version.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-openstack-version.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/sync-configuration-repository.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade-ceph.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade-manager.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade-services.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/wrapper-gilt.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/bootstrap/",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap/000-netbox.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap/300-openstack.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap/302-openstack-k8s-clusterapi-images.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/check/",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/100-ceph-with-ansible.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/100-ceph-with-rook.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/200-infrastructure.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/300-openstack.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/301-openstack-domain-manager.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/301-openstack-scs-standards.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/302-openstack-scs-compatible.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/302-openstack-with-tempest.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/303-prometheus-alert-status.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/500-kubernetes.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/customisations/",
                                "<f..t...... src/github.com/osism/testbed/scripts/customisations/external-api.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/deploy/",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/000-manager.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/001-helpers.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/100-ceph-with-ansible.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/100-ceph-with-rook.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/200-infrastructure.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/300-openstack.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/310-openstack-extended.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/320-openstack-minimal.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/400-monitoring.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/500-kubernetes.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/510-clusterapi.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/upgrade/",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/100-ceph-with-ansible.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/100-ceph-with-rook.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/200-infrastructure.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/300-openstack.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/310-openstack-extended.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/320-openstack-minimal.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/400-monitoring.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/500-kubernetes.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/510-clusterapi.sh",
                                ".d..t...... src/github.com/osism/testbed/terraform/",
                                "<f..t...... src/github.com/osism/testbed/terraform/Makefile",
                                "<f..t...... src/github.com/osism/testbed/terraform/README.md",
                                "<f..t...... src/github.com/osism/testbed/terraform/backend.env.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/backend.tf.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/clouds-public.yaml",
                                "<f..t...... src/github.com/osism/testbed/terraform/clouds.yaml.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/local.env.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/secure.yaml.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/terraformrc.sample",
                                ".d..t...... src/github.com/osism/testbed/terraform/environments/",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/artcodix.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/cah-f1a.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ci-centos-stream-9.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ci-debian-12.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ci-ubuntu-24.04.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ci.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/cleura.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/cloudpod.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/default.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/exigo.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/gx-scs.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/huaweicloud.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/otc.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ovh.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/pluscloudopen.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/regiocloud.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/wavestack.tfvars",
                                ".d..t...... src/github.com/osism/testbed/terraform/scripts/",
                                "<f..t...... src/github.com/osism/testbed/terraform/scripts/cleanup.py",
                                "*deleting   src/github.com/osism/testbed/ansible/ansible-manager-part-0.log",
                                "*deleting   src/github.com/osism/testbed/ansible/.id_rsa.ci.pub",
                                "*deleting   src/github.com/osism/testbed/ansible/.id_rsa.ci",
                                "*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/ci/terraform.tfstate",
                                "*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/ci/",
                                "*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_use_ephemeral_storage_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_boot_from_volume_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_boot_from_image_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_router_enable_snat_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_availability_zone_hints_router_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_availability_zone_hints_network_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/manager_boot_from_volume_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/manager_boot_from_image_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/neutron_floatingip_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/external_api_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/default_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/access_ipv6_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/access_ipv4_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/access_floatingip_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/terraform-provider-openstack_v3.4.0",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/README.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/LICENSE",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/CHANGELOG.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64.lock",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/terraform-provider-null",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/README.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/LICENSE",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/CHANGELOG.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64.lock",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/terraform-provider-local",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/README.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/LICENSE",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/CHANGELOG.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64.lock",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/environment",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/",
                                "*deleting   src/github.com/osism/testbed/terraform/variables.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/provider.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/outputs.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/nova.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/nodes_boot_from_volume_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/nodes.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/neutron_floatingip_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/neutron_availability_zone_hints_router_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/neutron_availability_zone_hints_network_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/neutron.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/manager_boot_from_volume_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/manager.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/main.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/local.env",
                                "*deleting   src/github.com/osism/testbed/terraform/inventory.ci",
                                "*deleting   src/github.com/osism/testbed/terraform/default_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/data.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/ci.auto.tfvars",
                                "*deleting   src/github.com/osism/testbed/terraform/access_floatingip_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform.lock.hcl",
                                "*deleting   src/github.com/osism/testbed/terraform/.id_rsa.ci.pub",
                                "*deleting   src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "*deleting   src/github.com/osism/testbed/terraform/.deploy.ci",
                                "*deleting   src/github.com/osism/testbed/terraform/.MANAGER_ADDRESS.ci"
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-1af1-8b27-00000000002a",
                        "name": "prepare-workspace-log",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_0/github.com/osism/zuul-config/roles/prepare-workspace-log"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:02:00.773553Z",
                            "start": "2026-03-21T00:00:38.163190Z"
                        },
                        "id": "0242ac1f-6606-1af1-8b27-00000000002d",
                        "name": "Synchronize src repos to workspace directory."
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 9,
            "failures": 0,
            "ignored": 0,
            "ok": 26,
            "rescued": 0,
            "skipped": 3,
            "unreachable": 0
        }
    },
    "trusted": true
},
{
    "branch": "main",
    "index": "1",
    "phase": "pre",
    "playbook": "github.com/osism/testbed/playbooks/pre.yml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-03-21T00:02:09.193972Z",
                    "start": "2026-03-21T00:02:02.206027Z"
                },
                "id": "0242ac1f-6606-6114-f43e-000000000002",
                "name": "Pre play"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "cloud": "ci"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:02:02.275623Z",
                            "start": "2026-03-21T00:02:02.224521Z"
                        },
                        "id": "0242ac1f-6606-6114-f43e-000000000004",
                        "name": "Set cloud fact (Zuul deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "changed": false,
                            "false_condition": "'nodepool' not in hostvars[groups['all'][0]]",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:02:02.306675Z",
                            "start": "2026-03-21T00:02:02.282331Z"
                        },
                        "id": "0242ac1f-6606-6114-f43e-000000000005",
                        "name": "Set cloud fact (local deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                            "delta": "0:00:01.647328",
                            "end": "2026-03-21 00:02:04.779294",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed04/src/github.com/osism/testbed/terraform/scripts",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-6114-f43e-000000000006-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-21 00:02:03.131966",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "2026-03-21 00:02:03 - clean up servers\n2026-03-21 00:02:04 - clean up keypairs\n2026-03-21 00:02:04 - wait for servers to be gone\n2026-03-21 00:02:04 - clean up ports\n2026-03-21 00:02:04 - clean up volumes\n2026-03-21 00:02:04 - disconnect routers\n2026-03-21 00:02:04 - clean up subnets\n2026-03-21 00:02:04 - clean up networks\n2026-03-21 00:02:04 - clean up security groups\n2026-03-21 00:02:04 - clean up floating ips\n2026-03-21 00:02:04 - clean up routers",
                            "stdout_lines": [
                                "2026-03-21 00:02:03 - clean up servers",
                                "2026-03-21 00:02:04 - clean up keypairs",
                                "2026-03-21 00:02:04 - wait for servers to be gone",
                                "2026-03-21 00:02:04 - clean up ports",
                                "2026-03-21 00:02:04 - clean up volumes",
                                "2026-03-21 00:02:04 - disconnect routers",
                                "2026-03-21 00:02:04 - clean up subnets",
                                "2026-03-21 00:02:04 - clean up networks",
                                "2026-03-21 00:02:04 - clean up security groups",
                                "2026-03-21 00:02:04 - clean up floating ips",
                                "2026-03-21 00:02:04 - clean up routers"
                            ],
                            "zuul_log_id": "0242ac1f-6606-6114-f43e-000000000006-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:02:05.281721Z",
                            "start": "2026-03-21T00:02:02.340879Z"
                        },
                        "id": "0242ac1f-6606-6114-f43e-000000000006",
                        "name": "Clean the cloud environment"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.unarchive",
                            "changed": false,
                            "dest": "/home/zuul-testbed04",
                            "gid": 1005,
                            "group": "zuul-testbed04",
                            "handler": "ZipArchive",
                            "invocation": {
                                "module_args": {
                                    "attributes": null,
                                    "copy": true,
                                    "creates": null,
                                    "decrypt": true,
                                    "dest": "/home/zuul-testbed04",
                                    "exclude": [],
                                    "extra_opts": [],
                                    "group": null,
                                    "include": [
                                        "tofu"
                                    ],
                                    "io_buffer_size": 65536,
                                    "keep_newer": false,
                                    "list_files": false,
                                    "mode": null,
                                    "owner": null,
                                    "remote_src": true,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "https://github.com/opentofu/opentofu/releases/download/v1.11.5/tofu_1.11.5_linux_amd64.zip",
                                    "unsafe_writes": false,
                                    "validate_certs": true
                                }
                            },
                            "mode": "0755",
                            "owner": "zuul-testbed04",
                            "size": 4096,
                            "src": "/home/zuul-testbed04/.ansible/tmp/ansible-tmp-1774051325.3292394-13-218544843607078/tofu_1.11.5_linux_amd64kz1y0qtu.zip",
                            "state": "directory",
                            "uid": 1004
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:02:07.695782Z",
                            "start": "2026-03-21T00:02:05.289872Z"
                        },
                        "id": "0242ac1f-6606-6114-f43e-000000000007",
                        "name": "Extract tofu binary"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "rsync",
                                "-avz",
                                "/home/zuul-testbed04/src/github.com/osism/terraform-base/testbed-default/",
                                "/home/zuul-testbed04/src/github.com/osism/testbed/terraform"
                            ],
                            "delta": "0:00:00.055312",
                            "end": "2026-03-21 00:02:08.038398",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "rsync -avz /home/zuul-testbed04/src/github.com/osism/terraform-base/testbed-default/ /home/zuul-testbed04/src/github.com/osism/testbed/terraform",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-6114-f43e-000000000008-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-21 00:02:07.983086",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "sending incremental file list\ndata.tf\nmain.tf\nmanager.tf\nneutron.tf\nnodes.tf\nnova.tf\noutputs.tf\nprovider.tf\nvariables.tf\ncustomisations/\ncustomisations/access_floatingip_custom.tf\ncustomisations/access_ipv4_custom.tf\ncustomisations/access_ipv6_custom.tf\ncustomisations/default_custom.tf\ncustomisations/external_api_custom.tf\ncustomisations/neutron_floatingip_custom.tf\noverrides/\noverrides/manager_boot_from_image_override.tf\noverrides/manager_boot_from_volume_override.tf\noverrides/neutron_availability_zone_hints_network_override.tf\noverrides/neutron_availability_zone_hints_router_override.tf\noverrides/neutron_router_enable_snat_override.tf\noverrides/nodes_boot_from_image_override.tf\noverrides/nodes_boot_from_volume_override.tf\noverrides/nodes_use_ephemeral_storage_override.tf\n\nsent 6,928 bytes  received 470 bytes  14,796.00 bytes/sec\ntotal size is 26,532  speedup is 3.59",
                            "stdout_lines": [
                                "sending incremental file list",
                                "data.tf",
                                "main.tf",
                                "manager.tf",
                                "neutron.tf",
                                "nodes.tf",
                                "nova.tf",
                                "outputs.tf",
                                "provider.tf",
                                "variables.tf",
                                "customisations/",
                                "customisations/access_floatingip_custom.tf",
                                "customisations/access_ipv4_custom.tf",
                                "customisations/access_ipv6_custom.tf",
                                "customisations/default_custom.tf",
                                "customisations/external_api_custom.tf",
                                "customisations/neutron_floatingip_custom.tf",
                                "overrides/",
                                "overrides/manager_boot_from_image_override.tf",
                                "overrides/manager_boot_from_volume_override.tf",
                                "overrides/neutron_availability_zone_hints_network_override.tf",
                                "overrides/neutron_availability_zone_hints_router_override.tf",
                                "overrides/neutron_router_enable_snat_override.tf",
                                "overrides/nodes_boot_from_image_override.tf",
                                "overrides/nodes_boot_from_volume_override.tf",
                                "overrides/nodes_use_ephemeral_storage_override.tf",
                                "",
                                "sent 6,928 bytes  received 470 bytes  14,796.00 bytes/sec",
                                "total size is 26,532  speedup is 3.59"
                            ],
                            "zuul_log_id": "0242ac1f-6606-6114-f43e-000000000008-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:02:08.249504Z",
                            "start": "2026-03-21T00:02:07.730325Z"
                        },
                        "id": "0242ac1f-6606-6114-f43e-000000000008",
                        "name": "Sync terraform blueprint"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.template",
                            "changed": true,
                            "checksum": "965738f20bea98af0fa5e0e54f73a8903ea2deb0",
                            "dest": "/home/zuul-testbed04/src/github.com/osism/testbed/terraform/local.env",
                            "diff": [],
                            "gid": 1005,
                            "group": "zuul-testbed04",
                            "invocation": {
                                "module_args": {
                                    "_original_basename": "local.env.j2",
                                    "attributes": null,
                                    "backup": false,
                                    "checksum": "965738f20bea98af0fa5e0e54f73a8903ea2deb0",
                                    "content": null,
                                    "dest": "/home/zuul-testbed04/src/github.com/osism/testbed/terraform/local.env",
                                    "directory_mode": null,
                                    "follow": false,
                                    "force": true,
                                    "group": null,
                                    "local_follow": null,
                                    "mode": "0644",
                                    "owner": null,
                                    "remote_src": null,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "/home/zuul-testbed04/.ansible/tmp/ansible-tmp-1774051328.3212636-22-109313640762664/source",
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "md5sum": "8ae5d13c5a02137867f5b04822cf5b92",
                            "mode": "0644",
                            "owner": "zuul-testbed04",
                            "size": 42,
                            "src": "/home/zuul-testbed04/.ansible/tmp/ansible-tmp-1774051328.3212636-22-109313640762664/source",
                            "state": "file",
                            "uid": 1004
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:02:09.193972Z",
                            "start": "2026-03-21T00:02:08.266245Z"
                        },
                        "id": "0242ac1f-6606-6114-f43e-000000000009",
                        "name": "Create local.env file"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 3,
            "failures": 0,
            "ignored": 0,
            "ok": 5,
            "rescued": 0,
            "skipped": 1,
            "unreachable": 0
        }
    },
    "trusted": false
},
{
    "branch": "main",
    "index": "0",
    "phase": "run",
    "playbook": "github.com/osism/testbed/playbooks/deploy.yml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-03-21T01:35:34.829570Z",
                    "start": "2026-03-21T00:02:10.205011Z"
                },
                "id": "0242ac1f-6606-9900-9c98-000000000002",
                "name": "Deploy testbed"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "_ceph_version": "reef",
                            "action": "ansible.builtin.debug",
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:02:10.304369Z",
                            "start": "2026-03-21T00:02:10.222573Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-000000000004",
                        "name": "Print used ceph version"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "_openstack_version": "2024.2",
                            "action": "ansible.builtin.debug",
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:02:10.410691Z",
                            "start": "2026-03-21T00:02:10.314514Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-000000000005",
                        "name": "Print used openstack version"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "_manager_version": "latest",
                            "action": "ansible.builtin.debug",
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:02:10.491827Z",
                            "start": "2026-03-21T00:02:10.418489Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-000000000006",
                        "name": "Print used manager version"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "cloud": "ci",
                                "image_username": "ubuntu"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:02:10.596151Z",
                            "start": "2026-03-21T00:02:10.501444Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-000000000007",
                        "name": "Set facts (Zuul deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "changed": false,
                            "false_condition": "'nodepool' not in hostvars[groups['all'][0]]",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:02:10.632356Z",
                            "start": "2026-03-21T00:02:10.606093Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-000000000008",
                        "name": "Set facts (local deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "make",
                                "ENVIRONMENT=ci-ubuntu-24.04",
                                "CLOUD=ci",
                                "IMAGE_USERNAME=ubuntu",
                                "TERRAFORM=/home/zuul-testbed04/tofu",
                                "VERSION_CEPH=reef",
                                "VERSION_MANAGER=latest",
                                "VERSION_OPENSTACK=2024.2",
                                "TEMPEST=true",
                                "CEPH_STACK=ceph-ansible",
                                "create"
                            ],
                            "delta": "0:01:47.949721",
                            "end": "2026-03-21 00:03:59.445060",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "make\n  ENVIRONMENT=ci-ubuntu-24.04\n  CLOUD=ci\n  IMAGE_USERNAME=ubuntu\n  TERRAFORM=/home/zuul-testbed04/tofu\n  VERSION_CEPH=reef\n  VERSION_MANAGER=latest\n  VERSION_OPENSTACK=2024.2\n  TEMPEST=true\n  CEPH_STACK=ceph-ansible\n  create\n",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed04/src/github.com/osism/testbed/terraform",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-9900-9c98-000000000009-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-21 00:02:11.495339",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "\nInitializing the backend...\n\nInitializing provider plugins...\n- terraform.io/builtin/terraform is built in to OpenTofu\n- Finding latest version of hashicorp/null...\n- Finding terraform-provider-openstack/openstack versions matching \">= 1.53.0\"...\n- Finding hashicorp/local versions matching \">= 2.2.0\"...\n- Installing hashicorp/local v2.7.0...\n- Installed hashicorp/local v2.7.0 (signed, key ID 0C0AF313E5FD9F80)\n- Installing hashicorp/null v3.2.4...\n- Installed hashicorp/null v3.2.4 (signed, key ID 0C0AF313E5FD9F80)\n- Installing terraform-provider-openstack/openstack v3.4.0...\n- Installed terraform-provider-openstack/openstack v3.4.0 (signed, key ID 4F80527A391BEFD2)\n\nProviders are signed by their developers.\nIf you'd like to know more about provider signing, you can read about it here:\nhttps://opentofu.org/docs/cli/plugins/signing/\n\nOpenTofu has created a lock file .terraform.lock.hcl to record the provider\nselections it made above. Include this file in your version control repository\nso that OpenTofu can guarantee to make the same selections by default when\nyou run \"tofu init\" in the future.\n\nOpenTofu has been successfully initialized!\n\nYou may now begin working with OpenTofu. Try running \"tofu plan\" to see\nany changes that are required for your infrastructure. All OpenTofu commands\nshould now work.\n\nIf you ever set or change modules or backend configuration for OpenTofu,\nrerun this command to reinitialize your working directory. If you forget, other\ncommands will detect it and remind you to do so if necessary.\nCreated and switched to workspace \"ci\"!\n\nYou're now on a new, empty workspace. Workspaces isolate their state,\nso if you run \"tofu plan\" OpenTofu will not see any existing state\nfor this configuration.\nci.auto.tfvars\ndefault_custom.tf\ndata.openstack_networking_network_v2.public: Reading...\ndata.openstack_networking_network_v2.public: Read complete after 1s [id=e6be7364-bfd8-4de7-8120-8f41c69a139a]\n\nOpenTofu used the selected providers to generate the following execution\nplan. Resource actions are indicated with the following symbols:\n  + create\n <= read (data resources)\n\nOpenTofu will perform the following actions:\n\n  # data.openstack_images_image_v2.image will be read during apply\n  # (config refers to values not yet known)\n <= data \"openstack_images_image_v2\" \"image\" {\n      + checksum    = (known after apply)\n      + created_at  = (known after apply)\n      + file        = (known after apply)\n      + id          = (known after apply)\n      + metadata    = (known after apply)\n      + min_disk_gb = (known after apply)\n      + min_ram_mb  = (known after apply)\n      + most_recent = true\n      + name        = (known after apply)\n      + protected   = (known after apply)\n      + region      = (known after apply)\n      + schema      = (known after apply)\n      + size_bytes  = (known after apply)\n      + tags        = (known after apply)\n      + updated_at  = (known after apply)\n    }\n\n  # data.openstack_images_image_v2.image_node will be read during apply\n  # (config refers to values not yet known)\n <= data \"openstack_images_image_v2\" \"image_node\" {\n      + checksum    = (known after apply)\n      + created_at  = (known after apply)\n      + file        = (known after apply)\n      + id          = (known after apply)\n      + metadata    = (known after apply)\n      + min_disk_gb = (known after apply)\n      + min_ram_mb  = (known after apply)\n      + most_recent = true\n      + name        = (known after apply)\n      + protected   = (known after apply)\n      + region      = (known after apply)\n      + schema      = (known after apply)\n      + size_bytes  = (known after apply)\n      + tags        = (known after apply)\n      + updated_at  = (known after apply)\n    }\n\n  # local_file.MANAGER_ADDRESS will be created\n  + resource \"local_file\" \"MANAGER_ADDRESS\" {\n      + content              = (known after apply)\n      + content_base64sha256 = (known after apply)\n      + content_base64sha512 = (known after apply)\n      + content_md5          = (known after apply)\n      + content_sha1         = (known after apply)\n      + content_sha256       = (known after apply)\n      + content_sha512       = (known after apply)\n      + directory_permission = \"0777\"\n      + file_permission      = \"0644\"\n      + filename             = \".MANAGER_ADDRESS.ci\"\n      + id                   = (known after apply)\n    }\n\n  # local_file.id_rsa_pub will be created\n  + resource \"local_file\" \"id_rsa_pub\" {\n      + content              = (known after apply)\n      + content_base64sha256 = (known after apply)\n      + content_base64sha512 = (known after apply)\n      + content_md5          = (known after apply)\n      + content_sha1         = (known after apply)\n      + content_sha256       = (known after apply)\n      + content_sha512       = (known after apply)\n      + directory_permission = \"0777\"\n      + file_permission      = \"0644\"\n      + filename             = \".id_rsa.ci.pub\"\n      + id                   = (known after apply)\n    }\n\n  # local_file.inventory will be created\n  + resource \"local_file\" \"inventory\" {\n      + content              = (known after apply)\n      + content_base64sha256 = (known after apply)\n      + content_base64sha512 = (known after apply)\n      + content_md5          = (known after apply)\n      + content_sha1         = (known after apply)\n      + content_sha256       = (known after apply)\n      + content_sha512       = (known after apply)\n      + directory_permission = \"0777\"\n      + file_permission      = \"0644\"\n      + filename             = \"inventory.ci\"\n      + id                   = (known after apply)\n    }\n\n  # local_sensitive_file.id_rsa will be created\n  + resource \"local_sensitive_file\" \"id_rsa\" {\n      + content              = (sensitive value)\n      + content_base64sha256 = (known after apply)\n      + content_base64sha512 = (known after apply)\n      + content_md5          = (known after apply)\n      + content_sha1         = (known after apply)\n      + content_sha256       = (known after apply)\n      + content_sha512       = (known after apply)\n      + directory_permission = \"0700\"\n      + file_permission      = \"0600\"\n      + filename             = \".id_rsa.ci\"\n      + id                   = (known after apply)\n    }\n\n  # null_resource.node_semaphore will be created\n  + resource \"null_resource\" \"node_semaphore\" {\n      + id = (known after apply)\n    }\n\n  # openstack_blockstorage_volume_v3.manager_base_volume[0] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"manager_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-manager-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[0] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-0-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[1] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-1-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[2] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-2-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[3] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-3-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[4] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-4-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[5] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-5-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[0] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-0-node-3\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[1] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-1-node-4\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[2] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-2-node-5\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[3] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-3-node-3\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[4] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-4-node-4\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[5] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-5-node-5\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[6] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-6-node-3\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[7] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-7-node-4\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[8] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-8-node-5\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_compute_instance_v2.manager_server will be created\n  + resource \"openstack_compute_instance_v2\" \"manager_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-4V-16\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-manager\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = (sensitive value)\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[0] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-0\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[1] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-1\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[2] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-2\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[3] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-3\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[4] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-4\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[5] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-5\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_keypair_v2.key will be created\n  + resource \"openstack_compute_keypair_v2\" \"key\" {\n      + fingerprint = (known after apply)\n      + id          = (known after apply)\n      + name        = \"testbed\"\n      + private_key = (sensitive value)\n      + public_key  = (known after apply)\n      + region      = (known after apply)\n      + user_id     = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[0] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[1] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[2] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[3] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[4] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[5] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[6] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[7] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[8] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_networking_floatingip_associate_v2.manager_floating_ip_association will be created\n  + resource \"openstack_networking_floatingip_associate_v2\" \"manager_floating_ip_association\" {\n      + fixed_ip    = (known after apply)\n      + floating_ip = (known after apply)\n      + id          = (known after apply)\n      + port_id     = (known after apply)\n      + region      = (known after apply)\n    }\n\n  # openstack_networking_floatingip_v2.manager_floating_ip will be created\n  + resource \"openstack_networking_floatingip_v2\" \"manager_floating_ip\" {\n      + address    = (known after apply)\n      + all_tags   = (known after apply)\n      + dns_domain = (known after apply)\n      + dns_name   = (known after apply)\n      + fixed_ip   = (known after apply)\n      + id         = (known after apply)\n      + pool       = \"public\"\n      + port_id    = (known after apply)\n      + region     = (known after apply)\n      + subnet_id  = (known after apply)\n      + tenant_id  = (known after apply)\n    }\n\n  # openstack_networking_network_v2.net_management will be created\n  + resource \"openstack_networking_network_v2\" \"net_management\" {\n      + admin_state_up          = (known after apply)\n      + all_tags                = (known after apply)\n      + availability_zone_hints = [\n          + \"nova\",\n        ]\n      + dns_domain              = (known after apply)\n      + external                = (known after apply)\n      + id                      = (known after apply)\n      + mtu                     = (known after apply)\n      + name                    = \"net-testbed-management\"\n      + port_security_enabled   = (known after apply)\n      + qos_policy_id           = (known after apply)\n      + region                  = (known after apply)\n      + shared                  = (known after apply)\n      + tenant_id               = (known after apply)\n      + transparent_vlan        = (known after apply)\n\n      + segments (known after apply)\n    }\n\n  # openstack_networking_port_v2.manager_port_management will be created\n  + resource \"openstack_networking_port_v2\" \"manager_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.5\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[0] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.10\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[1] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.11\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[2] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.12\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[3] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.13\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[4] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.14\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[5] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.15\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_router_interface_v2.router_interface will be created\n  + resource \"openstack_networking_router_interface_v2\" \"router_interface\" {\n      + force_destroy = false\n      + id            = (known after apply)\n      + port_id       = (known after apply)\n      + region        = (known after apply)\n      + router_id     = (known after apply)\n      + subnet_id     = (known after apply)\n    }\n\n  # openstack_networking_router_v2.router will be created\n  + resource \"openstack_networking_router_v2\" \"router\" {\n      + admin_state_up          = (known after apply)\n      + all_tags                = (known after apply)\n      + availability_zone_hints = [\n          + \"nova\",\n        ]\n      + distributed             = (known after apply)\n      + enable_snat             = (known after apply)\n      + external_network_id     = \"e6be7364-bfd8-4de7-8120-8f41c69a139a\"\n      + external_qos_policy_id  = (known after apply)\n      + id                      = (known after apply)\n      + name                    = \"testbed\"\n      + region                  = (known after apply)\n      + tenant_id               = (known after apply)\n\n      + external_fixed_ip (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule1 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule1\" {\n      + description             = \"ssh\"\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + port_range_max          = 22\n      + port_range_min          = 22\n      + protocol                = \"tcp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule2 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule2\" {\n      + description             = \"wireguard\"\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + port_range_max          = 51820\n      + port_range_min          = 51820\n      + protocol                = \"udp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule3 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule3\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"tcp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"192.168.16.0/20\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule4 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule4\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"udp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"192.168.16.0/20\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule5 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule5\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"icmp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_node_rule1 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule1\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"tcp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_node_rule2 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule2\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"udp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_node_rule3 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule3\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"icmp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_rule_vrrp will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_rule_vrrp\" {\n      + description             = \"vrrp\"\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"112\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_v2.security_group_management will be created\n  + resource \"openstack_networking_secgroup_v2\" \"security_group_management\" {\n      + all_tags    = (known after apply)\n      + description = \"management security group\"\n      + id          = (known after apply)\n      + name        = \"testbed-management\"\n      + region      = (known after apply)\n      + stateful    = (known after apply)\n      + tenant_id   = (known after apply)\n    }\n\n  # openstack_networking_secgroup_v2.security_group_node will be created\n  + resource \"openstack_networking_secgroup_v2\" \"security_group_node\" {\n      + all_tags    = (known after apply)\n      + description = \"node security group\"\n      + id          = (known after apply)\n      + name        = \"testbed-node\"\n      + region      = (known after apply)\n      + stateful    = (known after apply)\n      + tenant_id   = (known after apply)\n    }\n\n  # openstack_networking_subnet_v2.subnet_management will be created\n  + resource \"openstack_networking_subnet_v2\" \"subnet_management\" {\n      + all_tags          = (known after apply)\n      + cidr              = \"192.168.16.0/20\"\n      + dns_nameservers   = [\n          + \"8.8.8.8\",\n          + \"9.9.9.9\",\n        ]\n      + enable_dhcp       = true\n      + gateway_ip        = (known after apply)\n      + id                = (known after apply)\n      + ip_version        = 4\n      + ipv6_address_mode = (known after apply)\n      + ipv6_ra_mode      = (known after apply)\n      + name              = \"subnet-testbed-management\"\n      + network_id        = (known after apply)\n      + no_gateway        = false\n      + region            = (known after apply)\n      + service_types     = (known after apply)\n      + tenant_id         = (known after apply)\n\n      + allocation_pool {\n          + end   = \"192.168.31.250\"\n          + start = \"192.168.31.200\"\n        }\n    }\n\n  # terraform_data.image will be created\n  + resource \"terraform_data\" \"image\" {\n      + id     = (known after apply)\n      + input  = \"Ubuntu 24.04\"\n      + output = (known after apply)\n    }\n\n  # terraform_data.image_node will be created\n  + resource \"terraform_data\" \"image_node\" {\n      + id     = (known after apply)\n      + input  = \"Ubuntu 24.04\"\n      + output = (known after apply)\n    }\n\nPlan: 64 to add, 0 to change, 0 to destroy.\n\nChanges to Outputs:\n  + manager_address = (sensitive value)\n  + private_key     = (sensitive value)\nterraform_data.image: Creating...\nterraform_data.image: Creation complete after 0s [id=8008d7e4-bf62-c358-29be-2e969ff14724]\nterraform_data.image_node: Creating...\nterraform_data.image_node: Creation complete after 0s [id=e0a5da16-2dbd-e1c3-421b-a4369c546133]\ndata.openstack_images_image_v2.image_node: Reading...\ndata.openstack_images_image_v2.image: Reading...\nopenstack_networking_network_v2.net_management: Creating...\nopenstack_blockstorage_volume_v3.node_volume[0]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[5]: Creating...\nopenstack_compute_keypair_v2.key: Creating...\nopenstack_blockstorage_volume_v3.node_volume[4]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[3]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[7]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[8]: Creating...\ndata.openstack_images_image_v2.image: Read complete after 0s [id=846820b2-039e-4b42-adad-daf72e0f8ea4]\nopenstack_blockstorage_volume_v3.node_volume[1]: Creating...\ndata.openstack_images_image_v2.image_node: Read complete after 0s [id=846820b2-039e-4b42-adad-daf72e0f8ea4]\nopenstack_blockstorage_volume_v3.node_volume[2]: Creating...\nopenstack_compute_keypair_v2.key: Creation complete after 0s [id=testbed]\nopenstack_blockstorage_volume_v3.node_volume[6]: Creating...\nopenstack_networking_network_v2.net_management: Creation complete after 1s [id=7eaca87c-0a85-41bf-ab5e-6fdf6b5ced97]\nopenstack_blockstorage_volume_v3.manager_base_volume[0]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[3]: Creation complete after 4s [id=f08311eb-e2ad-4eb7-8da9-ac6d899a2409]\nopenstack_blockstorage_volume_v3.node_volume[5]: Creation complete after 4s [id=d7bb81ba-6491-4b5a-bb82-f0b49314eac0]\nopenstack_blockstorage_volume_v3.node_base_volume[0]: Creating...\nopenstack_blockstorage_volume_v3.node_base_volume[5]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[7]: Creation complete after 4s [id=4e74a544-8b3a-4a05-98a1-b7cc13f13c18]\nopenstack_blockstorage_volume_v3.node_base_volume[1]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[0]: Creation complete after 4s [id=e03b171b-b829-41a0-a0b6-cc46d85f013d]\nopenstack_blockstorage_volume_v3.node_base_volume[2]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[4]: Creation complete after 4s [id=666e73fe-87f5-4f65-9ff2-95a038c8b18d]\nopenstack_blockstorage_volume_v3.node_base_volume[4]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[8]: Creation complete after 4s [id=f76ae8f0-65cc-4a62-b19c-23f8a40fbde6]\nopenstack_blockstorage_volume_v3.node_base_volume[3]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[1]: Creation complete after 4s [id=8ae4aabb-0174-4999-a1d0-4bf78cbb2965]\nopenstack_blockstorage_volume_v3.node_volume[6]: Creation complete after 4s [id=9fa3c153-0e2a-45ef-8e48-e0095eb937ad]\nopenstack_blockstorage_volume_v3.node_volume[2]: Creation complete after 4s [id=cdffe5f1-7f0d-460e-a1db-d6846d226144]\nlocal_file.id_rsa_pub: Creating...\nopenstack_networking_subnet_v2.subnet_management: Creating...\nlocal_sensitive_file.id_rsa: Creating...\nlocal_file.id_rsa_pub: Creation complete after 0s [id=e9202c97db58affcf5ad9f31af35e04f072d79c9]\nlocal_sensitive_file.id_rsa: Creation complete after 0s [id=acb1728953243b6f3b98e025c3a88298a8c10704]\nopenstack_blockstorage_volume_v3.manager_base_volume[0]: Creation complete after 3s [id=651b152f-a2bc-4120-8c9d-c02531503cc8]\nopenstack_networking_subnet_v2.subnet_management: Creation complete after 1s [id=a0e6692c-16fe-4064-bb39-e6dfabcc46cb]\nopenstack_networking_router_v2.router: Creating...\nopenstack_blockstorage_volume_v3.node_base_volume[5]: Creation complete after 3s [id=95d0155a-2e41-44f5-9fae-3beacda427fd]\nopenstack_blockstorage_volume_v3.node_base_volume[0]: Creation complete after 3s [id=7cc5c56e-2215-4807-a75d-aa809849b26f]\nopenstack_blockstorage_volume_v3.node_base_volume[1]: Creation complete after 3s [id=64760a60-3a12-4fb7-90e5-3833dd4063b5]\nopenstack_blockstorage_volume_v3.node_base_volume[2]: Creation complete after 3s [id=d6e05184-e73c-49c9-a2ec-7c822155aa32]\nopenstack_blockstorage_volume_v3.node_base_volume[3]: Creation complete after 3s [id=b6130fcd-394f-4d1f-ac75-5973103f7ce0]\nopenstack_blockstorage_volume_v3.node_base_volume[4]: Creation complete after 3s [id=0fb4a7c9-731d-4ebf-9945-e69275d38ed5]\nopenstack_networking_router_v2.router: Creation complete after 3s [id=02497221-6595-4be2-b07d-622263862ec6]\nopenstack_networking_secgroup_v2.security_group_management: Creating...\nopenstack_networking_router_interface_v2.router_interface: Creating...\nopenstack_networking_secgroup_v2.security_group_node: Creating...\nopenstack_networking_secgroup_v2.security_group_node: Creation complete after 1s [id=bd22dc4e-182c-41d3-bf59-e4ff4cae3fd0]\nopenstack_networking_secgroup_rule_v2.security_group_node_rule3: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_node_rule1: Creating...\nopenstack_networking_port_v2.node_port_management[1]: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_rule_vrrp: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_node_rule2: Creating...\nopenstack_networking_port_v2.node_port_management[4]: Creating...\nopenstack_networking_port_v2.node_port_management[2]: Creating...\nopenstack_networking_port_v2.node_port_management[3]: Creating...\nopenstack_networking_secgroup_v2.security_group_management: Creation complete after 1s [id=78211e75-3b40-48d8-86f1-24ae0160c948]\nopenstack_networking_port_v2.node_port_management[5]: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_node_rule1: Creation complete after 0s [id=af496041-3bab-4302-bc3f-4298ee5e63ab]\nopenstack_networking_port_v2.node_port_management[0]: Creating...\nopenstack_networking_port_v2.node_port_management[3]: Creation complete after 1s [id=6d2237e8-6b42-425e-9866-5df0c0fde3a4]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule1: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_node_rule3: Creation complete after 1s [id=6f7ec8ea-a5fe-40d1-acc2-c71dc16ee44b]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule4: Creating...\nopenstack_networking_port_v2.node_port_management[1]: Creation complete after 1s [id=761114b3-7304-40c6-96a1-45962f98d109]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule5: Creating...\nopenstack_networking_port_v2.node_port_management[4]: Creation complete after 1s [id=526a6611-a846-4e7b-a6a1-0efffdc856e4]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule3: Creating...\nopenstack_networking_port_v2.node_port_management[5]: Creation complete after 1s [id=721c56ff-7560-4029-9a8f-7884ed9b5d65]\nopenstack_networking_port_v2.manager_port_management: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_rule_vrrp: Creation complete after 1s [id=358e6c81-45b8-4464-9a1a-78f462ac88ea]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule2: Creating...\nopenstack_networking_port_v2.node_port_management[2]: Creation complete after 1s [id=ec98f5b1-4c59-433a-a2fe-b28505829035]\nopenstack_networking_secgroup_rule_v2.security_group_node_rule2: Creation complete after 1s [id=34f1777d-d88e-4425-ab10-5dfcfce1679e]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule1: Creation complete after 0s [id=e7bc68ef-2317-495c-972b-2b0a2da673ae]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule4: Creation complete after 1s [id=eb5ac9c6-9466-49cb-a6ac-04fe575f50c6]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule5: Creation complete after 1s [id=6d6cbbf8-527e-4149-a845-509ae4e2bc5d]\nopenstack_networking_port_v2.manager_port_management: Creation complete after 1s [id=3f0d3a54-e327-4a03-b440-d8c037e98238]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule3: Creation complete after 1s [id=87dc9425-176e-4ffe-86a9-53435746cc61]\nopenstack_networking_port_v2.node_port_management[0]: Creation complete after 2s [id=e3bbeb11-4568-4312-9667-3bd35e5b9212]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule2: Creation complete after 1s [id=02772e8b-b987-4ce7-81d6-5f5c6071ed66]\nopenstack_networking_router_interface_v2.router_interface: Creation complete after 6s [id=650af6aa-405c-4e5a-900c-7458754ae491]\nopenstack_networking_floatingip_v2.manager_floating_ip: Creating...\nopenstack_compute_instance_v2.node_server[5]: Creating...\nopenstack_compute_instance_v2.node_server[2]: Creating...\nopenstack_compute_instance_v2.node_server[3]: Creating...\nopenstack_compute_instance_v2.node_server[1]: Creating...\nopenstack_compute_instance_v2.node_server[0]: Creating...\nopenstack_compute_instance_v2.node_server[4]: Creating...\nopenstack_networking_floatingip_v2.manager_floating_ip: Creation complete after 2s [id=b111c9fc-6c6b-40d6-bfd2-48a849c9d6fa]\nopenstack_networking_floatingip_associate_v2.manager_floating_ip_association: Creating...\nlocal_file.inventory: Creating...\nlocal_file.MANAGER_ADDRESS: Creating...\nlocal_file.MANAGER_ADDRESS: Creation complete after 0s [id=a3c79af417c06fce4a5c199d534ca0888020db5d]\nlocal_file.inventory: Creation complete after 0s [id=1214571b3308e4c65efbe083c65a548e2e6d04a2]\nopenstack_networking_floatingip_associate_v2.manager_floating_ip_association: Creation complete after 2s [id=b111c9fc-6c6b-40d6-bfd2-48a849c9d6fa]\nopenstack_compute_instance_v2.node_server[5]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[2]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[3]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[0]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[4]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[1]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[5]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[2]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[3]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[4]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[0]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[1]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[2]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[5]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[1]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[3]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[4]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[0]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[5]: Still creating... [40s elapsed]\nopenstack_compute_instance_v2.node_server[2]: Still creating... [40s elapsed]\nopenstack_compute_instance_v2.node_server[0]: Still creating... [40s elapsed]\nopenstack_compute_instance_v2.node_server[1]: Still creating... [40s elapsed]\nopenstack_compute_instance_v2.node_server[3]: Still creating... [40s elapsed]\nopenstack_compute_instance_v2.node_server[4]: Still creating... [40s elapsed]\nopenstack_compute_instance_v2.node_server[4]: Creation complete after 41s [id=fb568fa1-6091-4bf0-8a10-d9f5b4e9bc81]\nopenstack_compute_instance_v2.node_server[0]: Creation complete after 41s [id=b5e678c6-ecb3-463d-b36a-d05e6cfd8ce6]\nopenstack_compute_instance_v2.node_server[2]: Still creating... [50s elapsed]\nopenstack_compute_instance_v2.node_server[1]: Still creating... [50s elapsed]\nopenstack_compute_instance_v2.node_server[5]: Still creating... [50s elapsed]\nopenstack_compute_instance_v2.node_server[3]: Still creating... [50s elapsed]\nopenstack_compute_instance_v2.node_server[2]: Creation complete after 51s [id=63b931eb-2718-4b86-a042-3868feca6932]\nopenstack_compute_instance_v2.node_server[1]: Creation complete after 52s [id=12d0df05-7323-4b31-9296-02af5e15003e]\nopenstack_compute_instance_v2.node_server[3]: Creation complete after 52s [id=9c36ca74-ec65-467e-ac74-2465c8ecb2f3]\nopenstack_compute_instance_v2.node_server[5]: Still creating... [1m0s elapsed]\nopenstack_compute_instance_v2.node_server[5]: Creation complete after 1m2s [id=bf92fb57-087f-4e1a-b8d0-2ca65f70b679]\nnull_resource.node_semaphore: Creating...\nnull_resource.node_semaphore: Creation complete after 0s [id=4409973314172130239]\nopenstack_compute_volume_attach_v2.node_volume_attachment[4]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[6]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[3]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[1]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[2]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[7]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[0]: Creating...\nopenstack_compute_instance_v2.manager_server: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[5]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[8]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[6]: Creation complete after 3s [id=9c36ca74-ec65-467e-ac74-2465c8ecb2f3/9fa3c153-0e2a-45ef-8e48-e0095eb937ad]\nopenstack_compute_volume_attach_v2.node_volume_attachment[4]: Creation complete after 3s [id=fb568fa1-6091-4bf0-8a10-d9f5b4e9bc81/666e73fe-87f5-4f65-9ff2-95a038c8b18d]\nopenstack_compute_volume_attach_v2.node_volume_attachment[1]: Creation complete after 4s [id=fb568fa1-6091-4bf0-8a10-d9f5b4e9bc81/8ae4aabb-0174-4999-a1d0-4bf78cbb2965]\nopenstack_compute_volume_attach_v2.node_volume_attachment[2]: Creation complete after 4s [id=bf92fb57-087f-4e1a-b8d0-2ca65f70b679/cdffe5f1-7f0d-460e-a1db-d6846d226144]\nopenstack_compute_volume_attach_v2.node_volume_attachment[7]: Creation complete after 10s [id=fb568fa1-6091-4bf0-8a10-d9f5b4e9bc81/4e74a544-8b3a-4a05-98a1-b7cc13f13c18]\nopenstack_compute_volume_attach_v2.node_volume_attachment[3]: Creation complete after 10s [id=9c36ca74-ec65-467e-ac74-2465c8ecb2f3/f08311eb-e2ad-4eb7-8da9-ac6d899a2409]\nopenstack_compute_volume_attach_v2.node_volume_attachment[0]: Creation complete after 10s [id=9c36ca74-ec65-467e-ac74-2465c8ecb2f3/e03b171b-b829-41a0-a0b6-cc46d85f013d]\nopenstack_compute_volume_attach_v2.node_volume_attachment[5]: Creation complete after 10s [id=bf92fb57-087f-4e1a-b8d0-2ca65f70b679/d7bb81ba-6491-4b5a-bb82-f0b49314eac0]\nopenstack_compute_instance_v2.manager_server: Still creating... [10s elapsed]\nopenstack_compute_volume_attach_v2.node_volume_attachment[8]: Still creating... [10s elapsed]\nopenstack_compute_volume_attach_v2.node_volume_attachment[8]: Creation complete after 10s [id=bf92fb57-087f-4e1a-b8d0-2ca65f70b679/f76ae8f0-65cc-4a62-b19c-23f8a40fbde6]\nopenstack_compute_instance_v2.manager_server: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.manager_server: Creation complete after 21s [id=effe46d4-09b0-49e4-bd01-975dfcefbc4b]\n\nApply complete! Resources: 64 added, 0 changed, 0 destroyed.\n\nOutputs:\n\nmanager_address = <sensitive>\nprivate_key = <sensitive>",
                            "stdout_lines": [
                                "",
                                "Initializing the backend...",
                                "",
                                "Initializing provider plugins...",
                                "- terraform.io/builtin/terraform is built in to OpenTofu",
                                "- Finding latest version of hashicorp/null...",
                                "- Finding terraform-provider-openstack/openstack versions matching \">= 1.53.0\"...",
                                "- Finding hashicorp/local versions matching \">= 2.2.0\"...",
                                "- Installing hashicorp/local v2.7.0...",
                                "- Installed hashicorp/local v2.7.0 (signed, key ID 0C0AF313E5FD9F80)",
                                "- Installing hashicorp/null v3.2.4...",
                                "- Installed hashicorp/null v3.2.4 (signed, key ID 0C0AF313E5FD9F80)",
                                "- Installing terraform-provider-openstack/openstack v3.4.0...",
                                "- Installed terraform-provider-openstack/openstack v3.4.0 (signed, key ID 4F80527A391BEFD2)",
                                "",
                                "Providers are signed by their developers.",
                                "If you'd like to know more about provider signing, you can read about it here:",
                                "https://opentofu.org/docs/cli/plugins/signing/",
                                "",
                                "OpenTofu has created a lock file .terraform.lock.hcl to record the provider",
                                "selections it made above. Include this file in your version control repository",
                                "so that OpenTofu can guarantee to make the same selections by default when",
                                "you run \"tofu init\" in the future.",
                                "",
                                "OpenTofu has been successfully initialized!",
                                "",
                                "You may now begin working with OpenTofu. Try running \"tofu plan\" to see",
                                "any changes that are required for your infrastructure. All OpenTofu commands",
                                "should now work.",
                                "",
                                "If you ever set or change modules or backend configuration for OpenTofu,",
                                "rerun this command to reinitialize your working directory. If you forget, other",
                                "commands will detect it and remind you to do so if necessary.",
                                "Created and switched to workspace \"ci\"!",
                                "",
                                "You're now on a new, empty workspace. Workspaces isolate their state,",
                                "so if you run \"tofu plan\" OpenTofu will not see any existing state",
                                "for this configuration.",
                                "ci.auto.tfvars",
                                "default_custom.tf",
                                "data.openstack_networking_network_v2.public: Reading...",
                                "data.openstack_networking_network_v2.public: Read complete after 1s [id=e6be7364-bfd8-4de7-8120-8f41c69a139a]",
                                "",
                                "OpenTofu used the selected providers to generate the following execution",
                                "plan. Resource actions are indicated with the following symbols:",
                                "  + create",
                                " <= read (data resources)",
                                "",
                                "OpenTofu will perform the following actions:",
                                "",
                                "  # data.openstack_images_image_v2.image will be read during apply",
                                "  # (config refers to values not yet known)",
                                " <= data \"openstack_images_image_v2\" \"image\" {",
                                "      + checksum    = (known after apply)",
                                "      + created_at  = (known after apply)",
                                "      + file        = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + metadata    = (known after apply)",
                                "      + min_disk_gb = (known after apply)",
                                "      + min_ram_mb  = (known after apply)",
                                "      + most_recent = true",
                                "      + name        = (known after apply)",
                                "      + protected   = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + schema      = (known after apply)",
                                "      + size_bytes  = (known after apply)",
                                "      + tags        = (known after apply)",
                                "      + updated_at  = (known after apply)",
                                "    }",
                                "",
                                "  # data.openstack_images_image_v2.image_node will be read during apply",
                                "  # (config refers to values not yet known)",
                                " <= data \"openstack_images_image_v2\" \"image_node\" {",
                                "      + checksum    = (known after apply)",
                                "      + created_at  = (known after apply)",
                                "      + file        = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + metadata    = (known after apply)",
                                "      + min_disk_gb = (known after apply)",
                                "      + min_ram_mb  = (known after apply)",
                                "      + most_recent = true",
                                "      + name        = (known after apply)",
                                "      + protected   = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + schema      = (known after apply)",
                                "      + size_bytes  = (known after apply)",
                                "      + tags        = (known after apply)",
                                "      + updated_at  = (known after apply)",
                                "    }",
                                "",
                                "  # local_file.MANAGER_ADDRESS will be created",
                                "  + resource \"local_file\" \"MANAGER_ADDRESS\" {",
                                "      + content              = (known after apply)",
                                "      + content_base64sha256 = (known after apply)",
                                "      + content_base64sha512 = (known after apply)",
                                "      + content_md5          = (known after apply)",
                                "      + content_sha1         = (known after apply)",
                                "      + content_sha256       = (known after apply)",
                                "      + content_sha512       = (known after apply)",
                                "      + directory_permission = \"0777\"",
                                "      + file_permission      = \"0644\"",
                                "      + filename             = \".MANAGER_ADDRESS.ci\"",
                                "      + id                   = (known after apply)",
                                "    }",
                                "",
                                "  # local_file.id_rsa_pub will be created",
                                "  + resource \"local_file\" \"id_rsa_pub\" {",
                                "      + content              = (known after apply)",
                                "      + content_base64sha256 = (known after apply)",
                                "      + content_base64sha512 = (known after apply)",
                                "      + content_md5          = (known after apply)",
                                "      + content_sha1         = (known after apply)",
                                "      + content_sha256       = (known after apply)",
                                "      + content_sha512       = (known after apply)",
                                "      + directory_permission = \"0777\"",
                                "      + file_permission      = \"0644\"",
                                "      + filename             = \".id_rsa.ci.pub\"",
                                "      + id                   = (known after apply)",
                                "    }",
                                "",
                                "  # local_file.inventory will be created",
                                "  + resource \"local_file\" \"inventory\" {",
                                "      + content              = (known after apply)",
                                "      + content_base64sha256 = (known after apply)",
                                "      + content_base64sha512 = (known after apply)",
                                "      + content_md5          = (known after apply)",
                                "      + content_sha1         = (known after apply)",
                                "      + content_sha256       = (known after apply)",
                                "      + content_sha512       = (known after apply)",
                                "      + directory_permission = \"0777\"",
                                "      + file_permission      = \"0644\"",
                                "      + filename             = \"inventory.ci\"",
                                "      + id                   = (known after apply)",
                                "    }",
                                "",
                                "  # local_sensitive_file.id_rsa will be created",
                                "  + resource \"local_sensitive_file\" \"id_rsa\" {",
                                "      + content              = (sensitive value)",
                                "      + content_base64sha256 = (known after apply)",
                                "      + content_base64sha512 = (known after apply)",
                                "      + content_md5          = (known after apply)",
                                "      + content_sha1         = (known after apply)",
                                "      + content_sha256       = (known after apply)",
                                "      + content_sha512       = (known after apply)",
                                "      + directory_permission = \"0700\"",
                                "      + file_permission      = \"0600\"",
                                "      + filename             = \".id_rsa.ci\"",
                                "      + id                   = (known after apply)",
                                "    }",
                                "",
                                "  # null_resource.node_semaphore will be created",
                                "  + resource \"null_resource\" \"node_semaphore\" {",
                                "      + id = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.manager_base_volume[0] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"manager_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-manager-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[0] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-0-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[1] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-1-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[2] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-2-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[3] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-3-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[4] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-4-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[5] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-5-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[0] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-0-node-3\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[1] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-1-node-4\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[2] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-2-node-5\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[3] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-3-node-3\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[4] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-4-node-4\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[5] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-5-node-5\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[6] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-6-node-3\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[7] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-7-node-4\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[8] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-8-node-5\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.manager_server will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"manager_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-4V-16\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-manager\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = (sensitive value)",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[0] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-0\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[1] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-1\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[2] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-2\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[3] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-3\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[4] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-4\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[5] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-5\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_keypair_v2.key will be created",
                                "  + resource \"openstack_compute_keypair_v2\" \"key\" {",
                                "      + fingerprint = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + name        = \"testbed\"",
                                "      + private_key = (sensitive value)",
                                "      + public_key  = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + user_id     = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[0] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[1] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[2] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[3] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[4] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[5] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[6] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[7] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[8] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_floatingip_associate_v2.manager_floating_ip_association will be created",
                                "  + resource \"openstack_networking_floatingip_associate_v2\" \"manager_floating_ip_association\" {",
                                "      + fixed_ip    = (known after apply)",
                                "      + floating_ip = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + port_id     = (known after apply)",
                                "      + region      = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_floatingip_v2.manager_floating_ip will be created",
                                "  + resource \"openstack_networking_floatingip_v2\" \"manager_floating_ip\" {",
                                "      + address    = (known after apply)",
                                "      + all_tags   = (known after apply)",
                                "      + dns_domain = (known after apply)",
                                "      + dns_name   = (known after apply)",
                                "      + fixed_ip   = (known after apply)",
                                "      + id         = (known after apply)",
                                "      + pool       = \"public\"",
                                "      + port_id    = (known after apply)",
                                "      + region     = (known after apply)",
                                "      + subnet_id  = (known after apply)",
                                "      + tenant_id  = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_network_v2.net_management will be created",
                                "  + resource \"openstack_networking_network_v2\" \"net_management\" {",
                                "      + admin_state_up          = (known after apply)",
                                "      + all_tags                = (known after apply)",
                                "      + availability_zone_hints = [",
                                "          + \"nova\",",
                                "        ]",
                                "      + dns_domain              = (known after apply)",
                                "      + external                = (known after apply)",
                                "      + id                      = (known after apply)",
                                "      + mtu                     = (known after apply)",
                                "      + name                    = \"net-testbed-management\"",
                                "      + port_security_enabled   = (known after apply)",
                                "      + qos_policy_id           = (known after apply)",
                                "      + region                  = (known after apply)",
                                "      + shared                  = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "      + transparent_vlan        = (known after apply)",
                                "",
                                "      + segments (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.manager_port_management will be created",
                                "  + resource \"openstack_networking_port_v2\" \"manager_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.5\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[0] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.10\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[1] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.11\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[2] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.12\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[3] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.13\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[4] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.14\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[5] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.15\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_router_interface_v2.router_interface will be created",
                                "  + resource \"openstack_networking_router_interface_v2\" \"router_interface\" {",
                                "      + force_destroy = false",
                                "      + id            = (known after apply)",
                                "      + port_id       = (known after apply)",
                                "      + region        = (known after apply)",
                                "      + router_id     = (known after apply)",
                                "      + subnet_id     = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_router_v2.router will be created",
                                "  + resource \"openstack_networking_router_v2\" \"router\" {",
                                "      + admin_state_up          = (known after apply)",
                                "      + all_tags                = (known after apply)",
                                "      + availability_zone_hints = [",
                                "          + \"nova\",",
                                "        ]",
                                "      + distributed             = (known after apply)",
                                "      + enable_snat             = (known after apply)",
                                "      + external_network_id     = \"e6be7364-bfd8-4de7-8120-8f41c69a139a\"",
                                "      + external_qos_policy_id  = (known after apply)",
                                "      + id                      = (known after apply)",
                                "      + name                    = \"testbed\"",
                                "      + region                  = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "",
                                "      + external_fixed_ip (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule1 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule1\" {",
                                "      + description             = \"ssh\"",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + port_range_max          = 22",
                                "      + port_range_min          = 22",
                                "      + protocol                = \"tcp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule2 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule2\" {",
                                "      + description             = \"wireguard\"",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + port_range_max          = 51820",
                                "      + port_range_min          = 51820",
                                "      + protocol                = \"udp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule3 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule3\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"tcp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"192.168.16.0/20\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule4 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule4\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"udp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"192.168.16.0/20\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule5 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule5\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"icmp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_node_rule1 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule1\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"tcp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_node_rule2 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule2\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"udp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_node_rule3 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule3\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"icmp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_rule_vrrp will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_rule_vrrp\" {",
                                "      + description             = \"vrrp\"",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"112\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_v2.security_group_management will be created",
                                "  + resource \"openstack_networking_secgroup_v2\" \"security_group_management\" {",
                                "      + all_tags    = (known after apply)",
                                "      + description = \"management security group\"",
                                "      + id          = (known after apply)",
                                "      + name        = \"testbed-management\"",
                                "      + region      = (known after apply)",
                                "      + stateful    = (known after apply)",
                                "      + tenant_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_v2.security_group_node will be created",
                                "  + resource \"openstack_networking_secgroup_v2\" \"security_group_node\" {",
                                "      + all_tags    = (known after apply)",
                                "      + description = \"node security group\"",
                                "      + id          = (known after apply)",
                                "      + name        = \"testbed-node\"",
                                "      + region      = (known after apply)",
                                "      + stateful    = (known after apply)",
                                "      + tenant_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_subnet_v2.subnet_management will be created",
                                "  + resource \"openstack_networking_subnet_v2\" \"subnet_management\" {",
                                "      + all_tags          = (known after apply)",
                                "      + cidr              = \"192.168.16.0/20\"",
                                "      + dns_nameservers   = [",
                                "          + \"8.8.8.8\",",
                                "          + \"9.9.9.9\",",
                                "        ]",
                                "      + enable_dhcp       = true",
                                "      + gateway_ip        = (known after apply)",
                                "      + id                = (known after apply)",
                                "      + ip_version        = 4",
                                "      + ipv6_address_mode = (known after apply)",
                                "      + ipv6_ra_mode      = (known after apply)",
                                "      + name              = \"subnet-testbed-management\"",
                                "      + network_id        = (known after apply)",
                                "      + no_gateway        = false",
                                "      + region            = (known after apply)",
                                "      + service_types     = (known after apply)",
                                "      + tenant_id         = (known after apply)",
                                "",
                                "      + allocation_pool {",
                                "          + end   = \"192.168.31.250\"",
                                "          + start = \"192.168.31.200\"",
                                "        }",
                                "    }",
                                "",
                                "  # terraform_data.image will be created",
                                "  + resource \"terraform_data\" \"image\" {",
                                "      + id     = (known after apply)",
                                "      + input  = \"Ubuntu 24.04\"",
                                "      + output = (known after apply)",
                                "    }",
                                "",
                                "  # terraform_data.image_node will be created",
                                "  + resource \"terraform_data\" \"image_node\" {",
                                "      + id     = (known after apply)",
                                "      + input  = \"Ubuntu 24.04\"",
                                "      + output = (known after apply)",
                                "    }",
                                "",
                                "Plan: 64 to add, 0 to change, 0 to destroy.",
                                "",
                                "Changes to Outputs:",
                                "  + manager_address = (sensitive value)",
                                "  + private_key     = (sensitive value)",
                                "terraform_data.image: Creating...",
                                "terraform_data.image: Creation complete after 0s [id=8008d7e4-bf62-c358-29be-2e969ff14724]",
                                "terraform_data.image_node: Creating...",
                                "terraform_data.image_node: Creation complete after 0s [id=e0a5da16-2dbd-e1c3-421b-a4369c546133]",
                                "data.openstack_images_image_v2.image_node: Reading...",
                                "data.openstack_images_image_v2.image: Reading...",
                                "openstack_networking_network_v2.net_management: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[0]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[5]: Creating...",
                                "openstack_compute_keypair_v2.key: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[4]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[3]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[7]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[8]: Creating...",
                                "data.openstack_images_image_v2.image: Read complete after 0s [id=846820b2-039e-4b42-adad-daf72e0f8ea4]",
                                "openstack_blockstorage_volume_v3.node_volume[1]: Creating...",
                                "data.openstack_images_image_v2.image_node: Read complete after 0s [id=846820b2-039e-4b42-adad-daf72e0f8ea4]",
                                "openstack_blockstorage_volume_v3.node_volume[2]: Creating...",
                                "openstack_compute_keypair_v2.key: Creation complete after 0s [id=testbed]",
                                "openstack_blockstorage_volume_v3.node_volume[6]: Creating...",
                                "openstack_networking_network_v2.net_management: Creation complete after 1s [id=7eaca87c-0a85-41bf-ab5e-6fdf6b5ced97]",
                                "openstack_blockstorage_volume_v3.manager_base_volume[0]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[3]: Creation complete after 4s [id=f08311eb-e2ad-4eb7-8da9-ac6d899a2409]",
                                "openstack_blockstorage_volume_v3.node_volume[5]: Creation complete after 4s [id=d7bb81ba-6491-4b5a-bb82-f0b49314eac0]",
                                "openstack_blockstorage_volume_v3.node_base_volume[0]: Creating...",
                                "openstack_blockstorage_volume_v3.node_base_volume[5]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[7]: Creation complete after 4s [id=4e74a544-8b3a-4a05-98a1-b7cc13f13c18]",
                                "openstack_blockstorage_volume_v3.node_base_volume[1]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[0]: Creation complete after 4s [id=e03b171b-b829-41a0-a0b6-cc46d85f013d]",
                                "openstack_blockstorage_volume_v3.node_base_volume[2]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[4]: Creation complete after 4s [id=666e73fe-87f5-4f65-9ff2-95a038c8b18d]",
                                "openstack_blockstorage_volume_v3.node_base_volume[4]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[8]: Creation complete after 4s [id=f76ae8f0-65cc-4a62-b19c-23f8a40fbde6]",
                                "openstack_blockstorage_volume_v3.node_base_volume[3]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[1]: Creation complete after 4s [id=8ae4aabb-0174-4999-a1d0-4bf78cbb2965]",
                                "openstack_blockstorage_volume_v3.node_volume[6]: Creation complete after 4s [id=9fa3c153-0e2a-45ef-8e48-e0095eb937ad]",
                                "openstack_blockstorage_volume_v3.node_volume[2]: Creation complete after 4s [id=cdffe5f1-7f0d-460e-a1db-d6846d226144]",
                                "local_file.id_rsa_pub: Creating...",
                                "openstack_networking_subnet_v2.subnet_management: Creating...",
                                "local_sensitive_file.id_rsa: Creating...",
                                "local_file.id_rsa_pub: Creation complete after 0s [id=e9202c97db58affcf5ad9f31af35e04f072d79c9]",
                                "local_sensitive_file.id_rsa: Creation complete after 0s [id=acb1728953243b6f3b98e025c3a88298a8c10704]",
                                "openstack_blockstorage_volume_v3.manager_base_volume[0]: Creation complete after 3s [id=651b152f-a2bc-4120-8c9d-c02531503cc8]",
                                "openstack_networking_subnet_v2.subnet_management: Creation complete after 1s [id=a0e6692c-16fe-4064-bb39-e6dfabcc46cb]",
                                "openstack_networking_router_v2.router: Creating...",
                                "openstack_blockstorage_volume_v3.node_base_volume[5]: Creation complete after 3s [id=95d0155a-2e41-44f5-9fae-3beacda427fd]",
                                "openstack_blockstorage_volume_v3.node_base_volume[0]: Creation complete after 3s [id=7cc5c56e-2215-4807-a75d-aa809849b26f]",
                                "openstack_blockstorage_volume_v3.node_base_volume[1]: Creation complete after 3s [id=64760a60-3a12-4fb7-90e5-3833dd4063b5]",
                                "openstack_blockstorage_volume_v3.node_base_volume[2]: Creation complete after 3s [id=d6e05184-e73c-49c9-a2ec-7c822155aa32]",
                                "openstack_blockstorage_volume_v3.node_base_volume[3]: Creation complete after 3s [id=b6130fcd-394f-4d1f-ac75-5973103f7ce0]",
                                "openstack_blockstorage_volume_v3.node_base_volume[4]: Creation complete after 3s [id=0fb4a7c9-731d-4ebf-9945-e69275d38ed5]",
                                "openstack_networking_router_v2.router: Creation complete after 3s [id=02497221-6595-4be2-b07d-622263862ec6]",
                                "openstack_networking_secgroup_v2.security_group_management: Creating...",
                                "openstack_networking_router_interface_v2.router_interface: Creating...",
                                "openstack_networking_secgroup_v2.security_group_node: Creating...",
                                "openstack_networking_secgroup_v2.security_group_node: Creation complete after 1s [id=bd22dc4e-182c-41d3-bf59-e4ff4cae3fd0]",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule3: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule1: Creating...",
                                "openstack_networking_port_v2.node_port_management[1]: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_rule_vrrp: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule2: Creating...",
                                "openstack_networking_port_v2.node_port_management[4]: Creating...",
                                "openstack_networking_port_v2.node_port_management[2]: Creating...",
                                "openstack_networking_port_v2.node_port_management[3]: Creating...",
                                "openstack_networking_secgroup_v2.security_group_management: Creation complete after 1s [id=78211e75-3b40-48d8-86f1-24ae0160c948]",
                                "openstack_networking_port_v2.node_port_management[5]: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule1: Creation complete after 0s [id=af496041-3bab-4302-bc3f-4298ee5e63ab]",
                                "openstack_networking_port_v2.node_port_management[0]: Creating...",
                                "openstack_networking_port_v2.node_port_management[3]: Creation complete after 1s [id=6d2237e8-6b42-425e-9866-5df0c0fde3a4]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule1: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule3: Creation complete after 1s [id=6f7ec8ea-a5fe-40d1-acc2-c71dc16ee44b]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule4: Creating...",
                                "openstack_networking_port_v2.node_port_management[1]: Creation complete after 1s [id=761114b3-7304-40c6-96a1-45962f98d109]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule5: Creating...",
                                "openstack_networking_port_v2.node_port_management[4]: Creation complete after 1s [id=526a6611-a846-4e7b-a6a1-0efffdc856e4]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule3: Creating...",
                                "openstack_networking_port_v2.node_port_management[5]: Creation complete after 1s [id=721c56ff-7560-4029-9a8f-7884ed9b5d65]",
                                "openstack_networking_port_v2.manager_port_management: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_rule_vrrp: Creation complete after 1s [id=358e6c81-45b8-4464-9a1a-78f462ac88ea]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule2: Creating...",
                                "openstack_networking_port_v2.node_port_management[2]: Creation complete after 1s [id=ec98f5b1-4c59-433a-a2fe-b28505829035]",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule2: Creation complete after 1s [id=34f1777d-d88e-4425-ab10-5dfcfce1679e]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule1: Creation complete after 0s [id=e7bc68ef-2317-495c-972b-2b0a2da673ae]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule4: Creation complete after 1s [id=eb5ac9c6-9466-49cb-a6ac-04fe575f50c6]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule5: Creation complete after 1s [id=6d6cbbf8-527e-4149-a845-509ae4e2bc5d]",
                                "openstack_networking_port_v2.manager_port_management: Creation complete after 1s [id=3f0d3a54-e327-4a03-b440-d8c037e98238]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule3: Creation complete after 1s [id=87dc9425-176e-4ffe-86a9-53435746cc61]",
                                "openstack_networking_port_v2.node_port_management[0]: Creation complete after 2s [id=e3bbeb11-4568-4312-9667-3bd35e5b9212]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule2: Creation complete after 1s [id=02772e8b-b987-4ce7-81d6-5f5c6071ed66]",
                                "openstack_networking_router_interface_v2.router_interface: Creation complete after 6s [id=650af6aa-405c-4e5a-900c-7458754ae491]",
                                "openstack_networking_floatingip_v2.manager_floating_ip: Creating...",
                                "openstack_compute_instance_v2.node_server[5]: Creating...",
                                "openstack_compute_instance_v2.node_server[2]: Creating...",
                                "openstack_compute_instance_v2.node_server[3]: Creating...",
                                "openstack_compute_instance_v2.node_server[1]: Creating...",
                                "openstack_compute_instance_v2.node_server[0]: Creating...",
                                "openstack_compute_instance_v2.node_server[4]: Creating...",
                                "openstack_networking_floatingip_v2.manager_floating_ip: Creation complete after 2s [id=b111c9fc-6c6b-40d6-bfd2-48a849c9d6fa]",
                                "openstack_networking_floatingip_associate_v2.manager_floating_ip_association: Creating...",
                                "local_file.inventory: Creating...",
                                "local_file.MANAGER_ADDRESS: Creating...",
                                "local_file.MANAGER_ADDRESS: Creation complete after 0s [id=a3c79af417c06fce4a5c199d534ca0888020db5d]",
                                "local_file.inventory: Creation complete after 0s [id=1214571b3308e4c65efbe083c65a548e2e6d04a2]",
                                "openstack_networking_floatingip_associate_v2.manager_floating_ip_association: Creation complete after 2s [id=b111c9fc-6c6b-40d6-bfd2-48a849c9d6fa]",
                                "openstack_compute_instance_v2.node_server[5]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[2]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[3]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[0]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[4]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[1]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[5]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[2]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[3]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[4]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[0]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[1]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[2]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[5]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[1]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[3]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[4]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[0]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[5]: Still creating... [40s elapsed]",
                                "openstack_compute_instance_v2.node_server[2]: Still creating... [40s elapsed]",
                                "openstack_compute_instance_v2.node_server[0]: Still creating... [40s elapsed]",
                                "openstack_compute_instance_v2.node_server[1]: Still creating... [40s elapsed]",
                                "openstack_compute_instance_v2.node_server[3]: Still creating... [40s elapsed]",
                                "openstack_compute_instance_v2.node_server[4]: Still creating... [40s elapsed]",
                                "openstack_compute_instance_v2.node_server[4]: Creation complete after 41s [id=fb568fa1-6091-4bf0-8a10-d9f5b4e9bc81]",
                                "openstack_compute_instance_v2.node_server[0]: Creation complete after 41s [id=b5e678c6-ecb3-463d-b36a-d05e6cfd8ce6]",
                                "openstack_compute_instance_v2.node_server[2]: Still creating... [50s elapsed]",
                                "openstack_compute_instance_v2.node_server[1]: Still creating... [50s elapsed]",
                                "openstack_compute_instance_v2.node_server[5]: Still creating... [50s elapsed]",
                                "openstack_compute_instance_v2.node_server[3]: Still creating... [50s elapsed]",
                                "openstack_compute_instance_v2.node_server[2]: Creation complete after 51s [id=63b931eb-2718-4b86-a042-3868feca6932]",
                                "openstack_compute_instance_v2.node_server[1]: Creation complete after 52s [id=12d0df05-7323-4b31-9296-02af5e15003e]",
                                "openstack_compute_instance_v2.node_server[3]: Creation complete after 52s [id=9c36ca74-ec65-467e-ac74-2465c8ecb2f3]",
                                "openstack_compute_instance_v2.node_server[5]: Still creating... [1m0s elapsed]",
                                "openstack_compute_instance_v2.node_server[5]: Creation complete after 1m2s [id=bf92fb57-087f-4e1a-b8d0-2ca65f70b679]",
                                "null_resource.node_semaphore: Creating...",
                                "null_resource.node_semaphore: Creation complete after 0s [id=4409973314172130239]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[4]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[6]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[3]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[1]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[2]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[7]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[0]: Creating...",
                                "openstack_compute_instance_v2.manager_server: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[5]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[8]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[6]: Creation complete after 3s [id=9c36ca74-ec65-467e-ac74-2465c8ecb2f3/9fa3c153-0e2a-45ef-8e48-e0095eb937ad]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[4]: Creation complete after 3s [id=fb568fa1-6091-4bf0-8a10-d9f5b4e9bc81/666e73fe-87f5-4f65-9ff2-95a038c8b18d]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[1]: Creation complete after 4s [id=fb568fa1-6091-4bf0-8a10-d9f5b4e9bc81/8ae4aabb-0174-4999-a1d0-4bf78cbb2965]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[2]: Creation complete after 4s [id=bf92fb57-087f-4e1a-b8d0-2ca65f70b679/cdffe5f1-7f0d-460e-a1db-d6846d226144]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[7]: Creation complete after 10s [id=fb568fa1-6091-4bf0-8a10-d9f5b4e9bc81/4e74a544-8b3a-4a05-98a1-b7cc13f13c18]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[3]: Creation complete after 10s [id=9c36ca74-ec65-467e-ac74-2465c8ecb2f3/f08311eb-e2ad-4eb7-8da9-ac6d899a2409]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[0]: Creation complete after 10s [id=9c36ca74-ec65-467e-ac74-2465c8ecb2f3/e03b171b-b829-41a0-a0b6-cc46d85f013d]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[5]: Creation complete after 10s [id=bf92fb57-087f-4e1a-b8d0-2ca65f70b679/d7bb81ba-6491-4b5a-bb82-f0b49314eac0]",
                                "openstack_compute_instance_v2.manager_server: Still creating... [10s elapsed]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[8]: Still creating... [10s elapsed]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[8]: Creation complete after 10s [id=bf92fb57-087f-4e1a-b8d0-2ca65f70b679/f76ae8f0-65cc-4a62-b19c-23f8a40fbde6]",
                                "openstack_compute_instance_v2.manager_server: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.manager_server: Creation complete after 21s [id=effe46d4-09b0-49e4-bd01-975dfcefbc4b]",
                                "",
                                "Apply complete! Resources: 64 added, 0 changed, 0 destroyed.",
                                "",
                                "Outputs:",
                                "",
                                "manager_address = <sensitive>",
                                "private_key = <sensitive>"
                            ],
                            "zuul_log_id": "0242ac1f-6606-9900-9c98-000000000009-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:03:59.941897Z",
                            "start": "2026-03-21T00:02:10.671238Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-000000000009",
                        "name": "Create infrastructure (latest)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": false,
                            "false_condition": "_manager_version != \"latest\"",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:04:00.503565Z",
                            "start": "2026-03-21T00:03:59.992395Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-00000000000a",
                        "name": "Create infrastructure (stable)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.slurp",
                            "changed": false,
                            "content": "TUFOQUdFUl9BRERSRVNTPTgxLjE2My4xOTMuMTM2Cg==",
                            "encoding": "base64",
                            "invocation": {
                                "module_args": {
                                    "src": "/home/zuul-testbed04/src/github.com/osism/testbed/terraform/.MANAGER_ADDRESS.ci"
                                }
                            },
                            "source": "/home/zuul-testbed04/src/github.com/osism/testbed/terraform/.MANAGER_ADDRESS.ci"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:04:01.014777Z",
                            "start": "2026-03-21T00:04:00.528352Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-00000000000b",
                        "name": "Fetch manager address"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "manager_host": "81.163.193.136"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:04:01.104163Z",
                            "start": "2026-03-21T00:04:01.024646Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-00000000000c",
                        "name": "Set manager_host address"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "item",
                                    "changed": true,
                                    "cmd": [
                                        "~/venv/bin/ansible-galaxy",
                                        "collection",
                                        "install",
                                        "--force",
                                        "/home/zuul-testbed04/src/github.com/osism/ansible-collection-commons"
                                    ],
                                    "delta": "0:00:04.922437",
                                    "end": "2026-03-21 00:04:06.350208",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_raw_params": "~/venv/bin/ansible-galaxy collection install --force \"/home/zuul-testbed04/src/github.com/osism/ansible-collection-commons\"\n",
                                            "_uses_shell": false,
                                            "argv": null,
                                            "chdir": "/home/zuul-testbed04",
                                            "creates": null,
                                            "executable": null,
                                            "expand_argument_vars": true,
                                            "removes": null,
                                            "stdin": null,
                                            "stdin_add_newline": true,
                                            "strip_empty_ends": true,
                                            "zuul_ansible_split_streams": false,
                                            "zuul_log_id": "in-loop-ignore",
                                            "zuul_output_max_bytes": 1073741824
                                        }
                                    },
                                    "item": "commons",
                                    "msg": "",
                                    "rc": 0,
                                    "start": "2026-03-21 00:04:01.427771",
                                    "stderr": "",
                                    "stderr_lines": [],
                                    "stdout": "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2\n[WARNING]: Collection osism.services does not support Ansible version 2.15.2\nStarting galaxy collection install process\nProcess install dependency map\nStarting collection install process\nInstalling 'osism.commons:999.0.0' to '/home/zuul-testbed04/.ansible/collections/ansible_collections/osism/commons'\nCreated collection for osism.commons:999.0.0 at /home/zuul-testbed04/.ansible/collections/ansible_collections/osism/commons\nosism.commons:999.0.0 was installed successfully",
                                    "stdout_lines": [
                                        "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2",
                                        "[WARNING]: Collection osism.services does not support Ansible version 2.15.2",
                                        "Starting galaxy collection install process",
                                        "Process install dependency map",
                                        "Starting collection install process",
                                        "Installing 'osism.commons:999.0.0' to '/home/zuul-testbed04/.ansible/collections/ansible_collections/osism/commons'",
                                        "Created collection for osism.commons:999.0.0 at /home/zuul-testbed04/.ansible/collections/ansible_collections/osism/commons",
                                        "osism.commons:999.0.0 was installed successfully"
                                    ],
                                    "zuul_log_id": "in-loop-ignore"
                                },
                                {
                                    "ansible_loop_var": "item",
                                    "changed": true,
                                    "cmd": [
                                        "~/venv/bin/ansible-galaxy",
                                        "collection",
                                        "install",
                                        "--force",
                                        "/home/zuul-testbed04/src/github.com/osism/ansible-collection-services"
                                    ],
                                    "delta": "0:00:00.653376",
                                    "end": "2026-03-21 00:04:07.242512",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_raw_params": "~/venv/bin/ansible-galaxy collection install --force \"/home/zuul-testbed04/src/github.com/osism/ansible-collection-services\"\n",
                                            "_uses_shell": false,
                                            "argv": null,
                                            "chdir": "/home/zuul-testbed04",
                                            "creates": null,
                                            "executable": null,
                                            "expand_argument_vars": true,
                                            "removes": null,
                                            "stdin": null,
                                            "stdin_add_newline": true,
                                            "strip_empty_ends": true,
                                            "zuul_ansible_split_streams": false,
                                            "zuul_log_id": "in-loop-ignore",
                                            "zuul_output_max_bytes": 1073741824
                                        }
                                    },
                                    "item": "services",
                                    "msg": "",
                                    "rc": 0,
                                    "start": "2026-03-21 00:04:06.589136",
                                    "stderr": "",
                                    "stderr_lines": [],
                                    "stdout": "[WARNING]: Collection osism.services does not support Ansible version 2.15.2\n[WARNING]: Collection osism.commons does not support Ansible version 2.15.2\nStarting galaxy collection install process\nProcess install dependency map\nStarting collection install process\nInstalling 'osism.services:999.0.0' to '/home/zuul-testbed04/.ansible/collections/ansible_collections/osism/services'\nCreated collection for osism.services:999.0.0 at /home/zuul-testbed04/.ansible/collections/ansible_collections/osism/services\nosism.services:999.0.0 was installed successfully",
                                    "stdout_lines": [
                                        "[WARNING]: Collection osism.services does not support Ansible version 2.15.2",
                                        "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2",
                                        "Starting galaxy collection install process",
                                        "Process install dependency map",
                                        "Starting collection install process",
                                        "Installing 'osism.services:999.0.0' to '/home/zuul-testbed04/.ansible/collections/ansible_collections/osism/services'",
                                        "Created collection for osism.services:999.0.0 at /home/zuul-testbed04/.ansible/collections/ansible_collections/osism/services",
                                        "osism.services:999.0.0 was installed successfully"
                                    ],
                                    "zuul_log_id": "in-loop-ignore"
                                }
                            ]
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:04:07.295217Z",
                            "start": "2026-03-21T00:04:01.113283Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-00000000000d",
                        "name": "Update ansible collections"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.wait_for",
                            "changed": false,
                            "elapsed": 10,
                            "invocation": {
                                "module_args": {
                                    "active_connection_states": [
                                        "ESTABLISHED",
                                        "FIN_WAIT1",
                                        "FIN_WAIT2",
                                        "SYN_RECV",
                                        "SYN_SENT",
                                        "TIME_WAIT"
                                    ],
                                    "connect_timeout": 5,
                                    "delay": 10,
                                    "exclude_hosts": null,
                                    "host": "81.163.193.136",
                                    "msg": null,
                                    "path": null,
                                    "port": 22,
                                    "search_regex": "OpenSSH",
                                    "sleep": 1,
                                    "state": "started",
                                    "timeout": 300
                                }
                            },
                            "match_groupdict": {},
                            "match_groups": [],
                            "path": null,
                            "port": 22,
                            "search_regex": "OpenSSH",
                            "state": "started"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:04:17.843517Z",
                            "start": "2026-03-21T00:04:07.312762Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-00000000000e",
                        "name": "Wait up to 300 seconds for port 22 to become open and contain \"OpenSSH\""
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.pause",
                            "changed": false,
                            "delta": 60,
                            "echo": true,
                            "rc": 0,
                            "start": "2026-03-21 00:04:17.874921",
                            "stderr": "",
                            "stdout": "Paused for 60.0 seconds",
                            "stop": "2026-03-21 00:05:17.877335",
                            "user_input": ""
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:05:17.897054Z",
                            "start": "2026-03-21T00:04:17.851339Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-00000000000f",
                        "name": "Wait a little longer for the manager so that everything is ready"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
                            "changed": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:05:19.478819Z",
                            "start": "2026-03-21T00:05:17.968658Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-000000000010",
                        "name": "Fetch manager ssh hostkey"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "cp /home/zuul-testbed04/src/github.com/osism/testbed/terraform/.id* .",
                            "delta": "0:00:00.007418",
                            "end": "2026-03-21 00:05:19.899026",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "cp /home/zuul-testbed04/src/github.com/osism/testbed/terraform/.id* .",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed04/src/github.com/osism/testbed/ansible",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-9900-9c98-000000000011-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-21 00:05:19.891608",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "",
                            "stdout_lines": [],
                            "zuul_log_id": "0242ac1f-6606-9900-9c98-000000000011-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:05:20.034983Z",
                            "start": "2026-03-21T00:05:19.520331Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-000000000011",
                        "name": "Get ssh keypair from terraform environment"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.debug",
                            "changed": false,
                            "msg": "The task 'Run manager part 0' runs an Ansible playbook on the manager. There is no further output of this here. It takes a few minutes for this task to complete."
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:05:20.094756Z",
                            "start": "2026-03-21T00:05:20.052446Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-000000000012",
                        "name": "Point out that the following task takes some time and does not give any output"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "set -o pipefail\n~/venv/bin/ansible-playbook \\\n  -i /home/zuul-testbed04/src/github.com/osism/testbed/terraform/inventory.ci \\\n  --key-file /home/zuul-testbed04/src/github.com/osism/testbed/terraform/.id_rsa.ci \\\n  -e cloud=ci \\\n  -e terraform_environment=ci-ubuntu-24.04 \\\n  -e repo_path=/home/zuul-testbed04/src/github.com \\\n  manager-part-0.yml | tee -a ansible-manager-part-0.log\n",
                            "delta": "0:07:04.153215",
                            "end": "2026-03-21 00:12:24.652629",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "set -o pipefail\n~/venv/bin/ansible-playbook \\\n  -i /home/zuul-testbed04/src/github.com/osism/testbed/terraform/inventory.ci \\\n  --key-file /home/zuul-testbed04/src/github.com/osism/testbed/terraform/.id_rsa.ci \\\n  -e cloud=ci \\\n  -e terraform_environment=ci-ubuntu-24.04 \\\n  -e repo_path=/home/zuul-testbed04/src/github.com \\\n  manager-part-0.yml | tee -a ansible-manager-part-0.log\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed04/src/github.com/osism/testbed/ansible",
                                    "creates": null,
                                    "executable": "/bin/bash",
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-9900-9c98-000000000013-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-21 00:05:20.499414",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2\n\nPLAY [Wait for cloud-init to finish] *******************************************\n\nTASK [Check /var/lib/cloud/instance/boot-finished] *****************************\nok: [testbed-manager]\n\nPLAY [Run manager part 0] ******************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Get home directory of ansible user] **************************************\nok: [testbed-manager]\n\nTASK [Set repo_path fact] ******************************************************\nskipping: [testbed-manager]\n\nTASK [Update package cache] ****************************************************\nskipping: [testbed-manager]\n\nTASK [Install required packages] ***********************************************\nskipping: [testbed-manager]\n\nTASK [Remove some python packages] *********************************************\nskipping: [testbed-manager]\n\nTASK [Set venv_command fact (RedHat)] ******************************************\nskipping: [testbed-manager]\n\nTASK [Fail if Ubuntu version is lower than 24.04] ******************************\nskipping: [testbed-manager]\n\nTASK [Fail if Debian version is lower than 12] *********************************\nskipping: [testbed-manager]\n\nTASK [Set APT options on manager] **********************************************\nchanged: [testbed-manager]\n\nTASK [Update APT cache and run dist-upgrade] ***********************************\nchanged: [testbed-manager]\n\nTASK [Install HWE kernel package on Ubuntu] ************************************\nchanged: [testbed-manager]\n\nTASK [Install required packages] ***********************************************\nchanged: [testbed-manager]\n\nTASK [Remove some python packages] *********************************************\nchanged: [testbed-manager]\n\nTASK [Set venv_command fact (Debian)] ******************************************\nok: [testbed-manager]\n\nTASK [Get current user] ********************************************************\nok: [testbed-manager]\n\nTASK [Create venv directory] ***************************************************\nchanged: [testbed-manager]\n\nTASK [Install netaddr in venv] *************************************************\nchanged: [testbed-manager]\n\nTASK [Install ansible-core in venv] ********************************************\nchanged: [testbed-manager]\n\nTASK [Install requests >= 2.32.2] **********************************************\nchanged: [testbed-manager]\n\nTASK [Install docker >= 7.1.0] *************************************************\nchanged: [testbed-manager]\n\nTASK [Create directories in /opt/src] ******************************************\nchanged: [testbed-manager] => (item=osism/ansible-collection-commons)\nchanged: [testbed-manager] => (item=osism/ansible-collection-services)\n\nTASK [Sync sources in /opt/src] ************************************************\n[DEPRECATION WARNING]: The connection's stdin object is deprecated. Call \ndisplay.prompt_until(msg) instead. This feature will be removed in version \n2.19. Deprecation warnings can be disabled by setting \ndeprecation_warnings=False in ansible.cfg.\nchanged: [testbed-manager] => (item=osism/ansible-collection-commons)\nchanged: [testbed-manager] => (item=osism/ansible-collection-services)\n\nTASK [Create /usr/share/ansible directory] *************************************\nchanged: [testbed-manager]\n\nTASK [Install collections from Ansible galaxy] *********************************\nchanged: [testbed-manager] => (item=ansible.netcommon)\nchanged: [testbed-manager] => (item=ansible.posix)\nchanged: [testbed-manager] => (item=community.docker>=3.10.2)\n\nTASK [Install local collections] ***********************************************\nchanged: [testbed-manager] => (item=ansible-collection-commons)\nchanged: [testbed-manager] => (item=ansible-collection-services)\n\nPLAY [Create operator user] ****************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [osism.commons.operator : Gather variables for each operating system] *****\nok: [testbed-manager]\n\nTASK [osism.commons.operator : Set operator_groups variable to default value] ***\nok: [testbed-manager]\n\nTASK [osism.commons.operator : Create operator group] **************************\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Create user] ************************************\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Add user to additional groups] ******************\nchanged: [testbed-manager] => (item=adm)\nchanged: [testbed-manager] => (item=sudo)\n\nTASK [osism.commons.operator : Copy user sudoers file] *************************\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Set language variables in .bashrc configuration file] ***\nchanged: [testbed-manager] => (item=export LANGUAGE=C.UTF-8)\nchanged: [testbed-manager] => (item=export LANG=C.UTF-8)\nchanged: [testbed-manager] => (item=export LC_ALL=C.UTF-8)\n\nTASK [osism.commons.operator : Set custom environment variables in .bashrc configuration file] ***\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file] ***\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Create .ssh directory] **************************\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Check number of SSH authorized keys] ************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Set ssh authorized keys] ************************\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Delete ssh authorized keys] *********************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Set authorized GitHub accounts] *****************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Delete authorized GitHub accounts] **************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Set password] ***********************************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Unset & lock password] **************************\nok: [testbed-manager]\n\nPLAY [Run manager part 0] ******************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Recursively change ownership of /opt/venv] *******************************\nchanged: [testbed-manager]\n\nPLAY RECAP *********************************************************************\ntestbed-manager            : ok=33   changed=23   unreachable=0    failed=0    skipped=14   rescued=0    ignored=0   ",
                            "stdout_lines": [
                                "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2",
                                "",
                                "PLAY [Wait for cloud-init to finish] *******************************************",
                                "",
                                "TASK [Check /var/lib/cloud/instance/boot-finished] *****************************",
                                "ok: [testbed-manager]",
                                "",
                                "PLAY [Run manager part 0] ******************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Get home directory of ansible user] **************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Set repo_path fact] ******************************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Update package cache] ****************************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Install required packages] ***********************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Remove some python packages] *********************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Set venv_command fact (RedHat)] ******************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Fail if Ubuntu version is lower than 24.04] ******************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Fail if Debian version is lower than 12] *********************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Set APT options on manager] **********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Update APT cache and run dist-upgrade] ***********************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install HWE kernel package on Ubuntu] ************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install required packages] ***********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Remove some python packages] *********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Set venv_command fact (Debian)] ******************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Get current user] ********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Create venv directory] ***************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install netaddr in venv] *************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install ansible-core in venv] ********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install requests >= 2.32.2] **********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install docker >= 7.1.0] *************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Create directories in /opt/src] ******************************************",
                                "changed: [testbed-manager] => (item=osism/ansible-collection-commons)",
                                "changed: [testbed-manager] => (item=osism/ansible-collection-services)",
                                "",
                                "TASK [Sync sources in /opt/src] ************************************************",
                                "[DEPRECATION WARNING]: The connection's stdin object is deprecated. Call ",
                                "display.prompt_until(msg) instead. This feature will be removed in version ",
                                "2.19. Deprecation warnings can be disabled by setting ",
                                "deprecation_warnings=False in ansible.cfg.",
                                "changed: [testbed-manager] => (item=osism/ansible-collection-commons)",
                                "changed: [testbed-manager] => (item=osism/ansible-collection-services)",
                                "",
                                "TASK [Create /usr/share/ansible directory] *************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install collections from Ansible galaxy] *********************************",
                                "changed: [testbed-manager] => (item=ansible.netcommon)",
                                "changed: [testbed-manager] => (item=ansible.posix)",
                                "changed: [testbed-manager] => (item=community.docker>=3.10.2)",
                                "",
                                "TASK [Install local collections] ***********************************************",
                                "changed: [testbed-manager] => (item=ansible-collection-commons)",
                                "changed: [testbed-manager] => (item=ansible-collection-services)",
                                "",
                                "PLAY [Create operator user] ****************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Gather variables for each operating system] *****",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set operator_groups variable to default value] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Create operator group] **************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Create user] ************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Add user to additional groups] ******************",
                                "changed: [testbed-manager] => (item=adm)",
                                "changed: [testbed-manager] => (item=sudo)",
                                "",
                                "TASK [osism.commons.operator : Copy user sudoers file] *************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set language variables in .bashrc configuration file] ***",
                                "changed: [testbed-manager] => (item=export LANGUAGE=C.UTF-8)",
                                "changed: [testbed-manager] => (item=export LANG=C.UTF-8)",
                                "changed: [testbed-manager] => (item=export LC_ALL=C.UTF-8)",
                                "",
                                "TASK [osism.commons.operator : Set custom environment variables in .bashrc configuration file] ***",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file] ***",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Create .ssh directory] **************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Check number of SSH authorized keys] ************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set ssh authorized keys] ************************",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Delete ssh authorized keys] *********************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set authorized GitHub accounts] *****************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Delete authorized GitHub accounts] **************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set password] ***********************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Unset & lock password] **************************",
                                "ok: [testbed-manager]",
                                "",
                                "PLAY [Run manager part 0] ******************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Recursively change ownership of /opt/venv] *******************************",
                                "changed: [testbed-manager]",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "testbed-manager            : ok=33   changed=23   unreachable=0    failed=0    skipped=14   rescued=0    ignored=0   "
                            ],
                            "zuul_log_id": "0242ac1f-6606-9900-9c98-000000000013-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:12:24.716570Z",
                            "start": "2026-03-21T00:05:20.131785Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-000000000013",
                        "name": "Run manager part 0"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.debug",
                            "changed": false,
                            "msg": "It is now already possible to log in to the manager with 'make login'."
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:12:24.763196Z",
                            "start": "2026-03-21T00:12:24.728333Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-000000000014",
                        "name": "Point out that the log in on the manager is now possible"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.debug",
                            "changed": false,
                            "msg": "The task 'Run manager part 1 + 2' runs an Ansible playbook on the manager. There is no further output of this here. It takes a few minuts for this task to complete."
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:12:24.807792Z",
                            "start": "2026-03-21T00:12:24.773638Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-000000000015",
                        "name": "Point out that the following task takes some time and does not give any output"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "set -o pipefail\n~/venv/bin/ansible-playbook \\\n  -i /home/zuul-testbed04/src/github.com/osism/testbed/terraform/inventory.ci \\\n  --key-file /home/zuul-testbed04/src/github.com/osism/testbed/terraform/.id_rsa.ci \\\n  -e cloud=ci \\\n  -e terraform_environment=ci-ubuntu-24.04 \\\n  -e repo_path=/home/zuul-testbed04/src/github.com \\\n  -e manager_version=latest \\\n  manager-part-1.yml | tee -a ansible-manager-part-1.log\n",
                            "delta": "0:02:15.168434",
                            "end": "2026-03-21 00:14:40.340250",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "set -o pipefail\n~/venv/bin/ansible-playbook \\\n  -i /home/zuul-testbed04/src/github.com/osism/testbed/terraform/inventory.ci \\\n  --key-file /home/zuul-testbed04/src/github.com/osism/testbed/terraform/.id_rsa.ci \\\n  -e cloud=ci \\\n  -e terraform_environment=ci-ubuntu-24.04 \\\n  -e repo_path=/home/zuul-testbed04/src/github.com \\\n  -e manager_version=latest \\\n  manager-part-1.yml | tee -a ansible-manager-part-1.log\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed04/src/github.com/osism/testbed/ansible",
                                    "creates": null,
                                    "executable": "/bin/bash",
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-9900-9c98-000000000016-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-21 00:12:25.171816",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2\n\nPLAY [Run manager part 1] ******************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Set venv_command fact (RedHat)] ******************************************\nskipping: [testbed-manager]\n\nTASK [Set venv_command fact (Debian)] ******************************************\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Gather variables for each operating system] ***\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Set repository_default fact to default value] ***\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Set repositories to default] ******************\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Include distribution specific repository tasks] ***\nincluded: /home/zuul-testbed04/.ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-manager\n\nTASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************\nskipping: [testbed-manager]\n\nTASK [osism.commons.repository : Copy 99osism apt configuration] ***************\nchanged: [testbed-manager]\n\nTASK [osism.commons.repository : Remove sources.list file] *********************\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Copy ubuntu.sources file] *********************\nchanged: [testbed-manager]\n\nTASK [osism.commons.repository : Update package cache] *************************\nchanged: [testbed-manager]\n\nTASK [Get home directory of ansible user] **************************************\nok: [testbed-manager]\n\nTASK [Set repo_path fact] ******************************************************\nskipping: [testbed-manager]\n\nTASK [Copy SSH public key] *****************************************************\nchanged: [testbed-manager]\n\nTASK [Copy SSH private key] ****************************************************\nchanged: [testbed-manager]\n\nTASK [Create configuration directory] ******************************************\nchanged: [testbed-manager]\n\nTASK [Copy testbed repo] *******************************************************\n[DEPRECATION WARNING]: The connection's stdin object is deprecated. Call \ndisplay.prompt_until(msg) instead. This feature will be removed in version \n2.19. Deprecation warnings can be disabled by setting \ndeprecation_warnings=False in ansible.cfg.\nchanged: [testbed-manager]\n\nTASK [Install python requirements in venv] *************************************\nok: [testbed-manager] => (item=Jinja2)\nok: [testbed-manager] => (item=PyYAML)\nok: [testbed-manager] => (item=packaging)\nchanged: [testbed-manager] => (item=python-gilt==1.2.3)\nok: [testbed-manager] => (item=requests>=2.32.2)\nok: [testbed-manager] => (item=docker>=7.1.0)\n\nTASK [Copy testbed custom CA certificate on Debian/Ubuntu] *********************\nchanged: [testbed-manager]\n\nTASK [Copy testbed custom CA certificate on CentOS] ****************************\nskipping: [testbed-manager]\n\nTASK [Run update-ca-certificates on Debian/Ubuntu] *****************************\nchanged: [testbed-manager]\n\nTASK [Run update-ca-trust on RedHat] *******************************************\nskipping: [testbed-manager]\n\nTASK [Run manager part 2] ******************************************************\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.commons.repository : Force update of package cache] *****\nok: [testbed-manager]\n\nPLAY RECAP *********************************************************************\ntestbed-manager            : ok=21   changed=11   unreachable=0    failed=0    skipped=5    rescued=0    ignored=0   ",
                            "stdout_lines": [
                                "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2",
                                "",
                                "PLAY [Run manager part 1] ******************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Set venv_command fact (RedHat)] ******************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Set venv_command fact (Debian)] ******************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Gather variables for each operating system] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Set repository_default fact to default value] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Set repositories to default] ******************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Include distribution specific repository tasks] ***",
                                "included: /home/zuul-testbed04/.ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-manager",
                                "",
                                "TASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Copy 99osism apt configuration] ***************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Remove sources.list file] *********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Copy ubuntu.sources file] *********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Update package cache] *************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Get home directory of ansible user] **************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Set repo_path fact] ******************************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Copy SSH public key] *****************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Copy SSH private key] ****************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Create configuration directory] ******************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Copy testbed repo] *******************************************************",
                                "[DEPRECATION WARNING]: The connection's stdin object is deprecated. Call ",
                                "display.prompt_until(msg) instead. This feature will be removed in version ",
                                "2.19. Deprecation warnings can be disabled by setting ",
                                "deprecation_warnings=False in ansible.cfg.",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install python requirements in venv] *************************************",
                                "ok: [testbed-manager] => (item=Jinja2)",
                                "ok: [testbed-manager] => (item=PyYAML)",
                                "ok: [testbed-manager] => (item=packaging)",
                                "changed: [testbed-manager] => (item=python-gilt==1.2.3)",
                                "ok: [testbed-manager] => (item=requests>=2.32.2)",
                                "ok: [testbed-manager] => (item=docker>=7.1.0)",
                                "",
                                "TASK [Copy testbed custom CA certificate on Debian/Ubuntu] *********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Copy testbed custom CA certificate on CentOS] ****************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Run update-ca-certificates on Debian/Ubuntu] *****************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Run update-ca-trust on RedHat] *******************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Run manager part 2] ******************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.commons.repository : Force update of package cache] *****",
                                "ok: [testbed-manager]",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "testbed-manager            : ok=21   changed=11   unreachable=0    failed=0    skipped=5    rescued=0    ignored=0   "
                            ],
                            "zuul_log_id": "0242ac1f-6606-9900-9c98-000000000016-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:14:40.470617Z",
                            "start": "2026-03-21T00:12:24.857725Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-000000000016",
                        "name": "Run manager part 1 + 2"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed04/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.193.136",
                                "sudo",
                                "shutdown",
                                "-r",
                                "now"
                            ],
                            "delta": "0:00:01.016429",
                            "end": "2026-03-21 00:14:41.803988",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed04/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.193.136 sudo shutdown -r now",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-9900-9c98-000000000017-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-21 00:14:40.787559",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "",
                            "stdout_lines": [],
                            "zuul_log_id": "0242ac1f-6606-9900-9c98-000000000017-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:14:42.033028Z",
                            "start": "2026-03-21T00:14:40.520811Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-000000000017",
                        "name": "Reboot manager"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.wait_for",
                            "changed": false,
                            "elapsed": 14,
                            "invocation": {
                                "module_args": {
                                    "active_connection_states": [
                                        "ESTABLISHED",
                                        "FIN_WAIT1",
                                        "FIN_WAIT2",
                                        "SYN_RECV",
                                        "SYN_SENT",
                                        "TIME_WAIT"
                                    ],
                                    "connect_timeout": 5,
                                    "delay": 10,
                                    "exclude_hosts": null,
                                    "host": "81.163.193.136",
                                    "msg": null,
                                    "path": null,
                                    "port": 22,
                                    "search_regex": "OpenSSH",
                                    "sleep": 1,
                                    "state": "started",
                                    "timeout": 300
                                }
                            },
                            "match_groupdict": {},
                            "match_groups": [],
                            "path": null,
                            "port": 22,
                            "search_regex": "OpenSSH",
                            "state": "started"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:14:56.789769Z",
                            "start": "2026-03-21T00:14:42.047262Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-000000000018",
                        "name": "Wait up to 300 seconds for port 22 to become open and contain \"OpenSSH\""
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.pause",
                            "changed": false,
                            "delta": 60,
                            "echo": true,
                            "rc": 0,
                            "start": "2026-03-21 00:14:56.815072",
                            "stderr": "",
                            "stdout": "Paused for 60.0 seconds",
                            "stop": "2026-03-21 00:15:56.817119",
                            "user_input": ""
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:15:56.840539Z",
                            "start": "2026-03-21T00:14:56.798956Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-000000000019",
                        "name": "Wait a little longer for the manager so that everything is ready"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed04/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.193.136",
                                "/opt/configuration/scripts/deploy-manager.sh"
                            ],
                            "delta": "0:23:35.499488",
                            "end": "2026-03-21 00:39:33.047636",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed04/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.193.136 /opt/configuration/scripts/deploy-manager.sh",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-9900-9c98-00000000001a-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-21 00:15:57.548148",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "\n# DEPLOY MANAGER\n\n+ set -e\n+ echo\n+ echo '# DEPLOY MANAGER'\n+ echo\n+ cat /opt/manager-vars.sh\nexport NUMBER_OF_NODES=6\n\nexport CEPH_VERSION=reef\nexport CONFIGURATION_VERSION=main\nexport MANAGER_VERSION=latest\nexport OPENSTACK_VERSION=2024.2\n\nexport ARA=false\nexport DEPLOY_MODE=manager\nexport TEMPEST=true\nexport IS_ZUUL=true\n\nexport MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136\nexport EXTERNAL_API=false\n\nexport IMAGE_USER=ubuntu\nexport IMAGE_NODE_USER=ubuntu\n\nexport CEPH_STACK=ceph-ansible\n\n+ echo\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=true\n++ TEMPEST=true\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ sudo ln -sf /opt/configuration/contrib/semver2.sh /usr/local/bin/semver\n+ docker version\nClient: Docker Engine - Community\n Version:           27.5.1\n API version:       1.47\n Go version:        go1.22.11\n Git commit:        9f9e405\n Built:             Wed Jan 22 13:41:48 2025\n OS/Arch:           linux/amd64\n Context:           default\n\nServer: Docker Engine - Community\n Engine:\n  Version:          27.5.1\n  API version:      1.47 (minimum version 1.24)\n  Go version:       go1.22.11\n  Git commit:       4c9b3b0\n  Built:            Wed Jan 22 13:41:48 2025\n  OS/Arch:          linux/amd64\n  Experimental:     false\n containerd:\n  Version:          v2.2.2\n  GitCommit:        301b2dac98f15c27117da5c8af12118a041a31d9\n runc:\n  Version:          1.3.4\n  GitCommit:        v1.3.4-0-gd6d73eb8\n docker-init:\n  Version:          0.19.0\n  GitCommit:        de40ad0\n+ sh -c /opt/configuration/scripts/deploy/000-manager.sh\n+ set -e\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=true\n++ TEMPEST=true\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ /opt/configuration/scripts/set-ceph-version.sh reef\n+ set -e\n+ VERSION=reef\n++ grep '^ceph_version:' /opt/configuration/environments/manager/configuration.yml\n+ [[ -n ceph_version: reef ]]\n+ sed -i 's/ceph_version: .*/ceph_version: reef/g' /opt/configuration/environments/manager/configuration.yml\n+ /opt/configuration/scripts/set-openstack-version.sh 2024.2\n+ set -e\n+ VERSION=2024.2\n++ grep '^openstack_version:' /opt/configuration/environments/manager/configuration.yml\n+ [[ -n openstack_version: 2024.2 ]]\n+ sed -i 's/openstack_version: .*/openstack_version: 2024.2/g' /opt/configuration/environments/manager/configuration.yml\n+ [[ ceph-ansible == \\r\\o\\o\\k ]]\n++ semver latest 7.0.0\n+ [[ -1 -ge 0 ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ echo 'enable_osism_kubernetes: true'\n++ semver latest 10.0.0-0\n+ [[ -1 -ge 0 ]]\n++ semver 2024.2 2025.1\n+ [[ -1 -ge 0 ]]\n+ /opt/configuration/scripts/enable-resource-nodes.sh\n+ [[ -e /opt/venv/bin/activate ]]\n+ source /opt/venv/bin/activate\n++ deactivate nondestructive\n++ '[' -n '' ']'\n++ '[' -n '' ']'\n++ hash -r\n++ '[' -n '' ']'\n++ unset VIRTUAL_ENV\n++ unset VIRTUAL_ENV_PROMPT\n++ '[' '!' nondestructive = nondestructive ']'\n++ '[' linux-gnu = cygwin ']'\n++ '[' linux-gnu = msys ']'\n++ export VIRTUAL_ENV=/opt/venv\n++ VIRTUAL_ENV=/opt/venv\n++ _OLD_VIRTUAL_PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin\n++ PATH=/opt/venv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin\n++ export PATH\n++ '[' -n '' ']'\n++ '[' -z '' ']'\n++ _OLD_VIRTUAL_PS1=\n++ PS1='(venv) '\n++ export PS1\n++ VIRTUAL_ENV_PROMPT='(venv) '\n++ export VIRTUAL_ENV_PROMPT\n++ hash -r\n+ ansible-playbook -i testbed-manager, --vault-password-file /opt/configuration/environments/.vault_pass /opt/configuration/ansible/manager-part-3.yml\n\nPLAY [Copy custom facts] *******************************************************\n\nTASK [Create custom facts directory] *******************************************\nok: [testbed-manager]\n\nTASK [Copy fact files] *********************************************************\nchanged: [testbed-manager]\n\nPLAY [Before the deployment of the manager] ************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Get /opt/manager-vars.sh] ************************************************\nok: [testbed-manager]\n\nTASK [Add ara_server_mariadb_volume_type parameter] ****************************\nchanged: [testbed-manager]\n\nTASK [Add netbox_enable parameter] *********************************************\nskipping: [testbed-manager]\n\nTASK [Install HWE kernel package on Ubuntu] ************************************\nchanged: [testbed-manager]\n\nTASK [Check if /etc/OTC_region exist] ******************************************\nok: [testbed-manager]\n\nTASK [Add nova_compute_virt_type parameter] ************************************\nskipping: [testbed-manager]\n\nPLAY [Apply role traefik] ******************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Apply traefik role] ******************************************************\nincluded: osism.services.traefik for testbed-manager\n\nTASK [osism.services.traefik : Include config tasks] ***************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/traefik/tasks/config.yml for testbed-manager\n\nTASK [osism.services.traefik : Create required directories] ********************\nchanged: [testbed-manager] => (item=/opt/traefik)\nchanged: [testbed-manager] => (item=/opt/traefik/certificates)\nchanged: [testbed-manager] => (item=/opt/traefik/configuration)\n\nTASK [osism.services.traefik : Copy configuration files] ***********************\nchanged: [testbed-manager] => (item=traefik.yml)\nchanged: [testbed-manager] => (item=traefik.env)\nchanged: [testbed-manager] => (item=certificates.yml)\n\nTASK [osism.services.traefik : Copy certificate cert files] ********************\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager]\n\nTASK [osism.services.traefik : Copy certificate key files] *********************\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager]\n\nTASK [osism.services.traefik : Copy dynamic configuration] *********************\nskipping: [testbed-manager]\n\nTASK [osism.services.traefik : Remove dynamic configuration] *******************\nok: [testbed-manager]\n\nTASK [osism.services.traefik : Include service tasks] **************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/traefik/tasks/service.yml for testbed-manager\n\nTASK [osism.services.traefik : Create traefik external network] ****************\nchanged: [testbed-manager]\n\nTASK [osism.services.traefik : Copy docker-compose.yml file] *******************\nchanged: [testbed-manager]\n\nTASK [osism.services.traefik : Manage traefik service] *************************\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.services.traefik : Restart traefik service] *************\nskipping: [testbed-manager]\n\nPLAY [Deploy manager service] **************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Apply manager role] ******************************************************\nincluded: osism.services.manager for testbed-manager\n\nTASK [osism.services.manager : Include install tasks] **************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/install-Debian-family.yml for testbed-manager\n\nTASK [osism.services.manager : Install required packages] **********************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Gather variables for each operating system] *****\nok: [testbed-manager]\n\nTASK [osism.services.manager : Include config tasks] ***************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config.yml for testbed-manager\n\nTASK [osism.services.manager : Create required directories] ********************\nchanged: [testbed-manager] => (item=/opt/ansible)\nchanged: [testbed-manager] => (item=/opt/archive)\nchanged: [testbed-manager] => (item=/opt/manager/configuration)\nchanged: [testbed-manager] => (item=/opt/manager/data)\nok: [testbed-manager] => (item=/opt/manager)\nchanged: [testbed-manager] => (item=/opt/manager/secrets)\nchanged: [testbed-manager] => (item=/opt/ansible/secrets)\nchanged: [testbed-manager] => (item=/opt/state)\n\nTASK [osism.services.manager : Copy all environment file] **********************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy client environment file] *******************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include ara config tasks] ***********************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ara.yml for testbed-manager\n\nTASK [osism.services.manager : Copy ARA environment files] *********************\nchanged: [testbed-manager] => (item=ara)\nchanged: [testbed-manager] => (item=ara-server)\n\nTASK [osism.services.manager : Copy MariaDB environment file] ******************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include vault config tasks] *********************\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Include frontend config tasks] ******************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-frontend.yml for testbed-manager\n\nTASK [osism.services.manager : Copy frontend environment file] *****************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include ansible config tasks] *******************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ansible.yml for testbed-manager\n\nTASK [osism.services.manager : Copy private ssh keys] **************************\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy ansible environment file] ******************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include netbox config tasks] ********************\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Include celery config tasks] ********************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-celery.yml for testbed-manager\n\nTASK [osism.services.manager : Set fs.inotify.max_user_watches] ****************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Set fs.inotify.max_user_instances] **************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy celery environment files] ******************\nchanged: [testbed-manager] => (item=conductor)\nchanged: [testbed-manager] => (item=openstack)\n\nTASK [osism.services.manager : Copy listener environment file] *****************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Check for conductor.yml] ************************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy conductor configuration file] **************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy empty conductor configuration file] ********\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Include wrapper config tasks] *******************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-wrapper.yml for testbed-manager\n\nTASK [osism.services.manager : Include wrapper vars file] **********************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy wrapper scripts] ***************************\nchanged: [testbed-manager] => (item=osism)\nchanged: [testbed-manager] => (item=osism-update-docker)\nchanged: [testbed-manager] => (item=osism-update-manager)\n\nTASK [osism.services.manager : Copy cilium wrapper script] *********************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy hubble wrapper script] *********************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy flux wrapper script] ***********************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include scripts config tasks] *******************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-scripts.yml for testbed-manager\n\nTASK [osism.services.manager : Include scripts vars file] **********************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy scripts] ***********************************\nchanged: [testbed-manager] => (item=osism-include)\n\nTASK [osism.services.manager : Include service tasks] **************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/service.yml for testbed-manager\n\nTASK [osism.services.manager : Copy manager systemd unit file] *****************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Create traefik external network] ****************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Set mariadb healthcheck for mariadb < 11.0.0] ***\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Set mariadb healthcheck for mariadb >= 11.0.0] ***\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy docker-compose.yml file] *******************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Pull container images] **************************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Stop and disable old service docker-compose@manager] ***\nok: [testbed-manager]\n\nTASK [osism.services.manager : Do a manual start of the manager service] *******\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Manage manager service] *************************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Register that manager service was started] ******\nok: [testbed-manager]\n\nTASK [osism.services.manager : Flush handlers] *********************************\n\nRUNNING HANDLER [osism.services.manager : Restart manager service] *************\nskipping: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Wait for manager service to start] ***\nPausing for 60 seconds\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Ensure that all containers are up] ***\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Wait for an healthy manager service] ***\nFAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (50 retries left).\nFAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (49 retries left).\nFAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (48 retries left).\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Copy osismclient bash completion script] ***\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include initialize tasks] ***********************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/initialize.yml for testbed-manager\n\nTASK [osism.services.manager : Flush handlers] *********************************\n\nTASK [osism.services.manager : Include vault initialize tasks] *****************\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Include version verification tasks] *************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/verify-versions.yml for testbed-manager\n\nTASK [osism.services.manager : Deploy service manager version check script] ****\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Execute service manager version check] **********\nok: [testbed-manager]\n\nTASK [osism.services.manager : Display version check results] ******************\nok: [testbed-manager] => {\n    \"version_check_result.stdout_lines\": [\n        \"=== OSISM Container Version Check ===\",\n        \"Checking running containers against expected versions...\",\n        \"\",\n        \"Checking service: inventory_reconciler (Inventory Reconciler Service)\",\n        \"  Expected: registry.osism.tech/osism/inventory-reconciler:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/inventory-reconciler:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: osism-ansible (OSISM Ansible Service)\",\n        \"  Expected: registry.osism.tech/osism/osism-ansible:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism-ansible:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: osism-kubernetes (Osism-Kubernetes Service)\",\n        \"  Expected: registry.osism.tech/osism/osism-kubernetes:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism-kubernetes:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: ceph-ansible (Ceph-Ansible Service)\",\n        \"  Expected: registry.osism.tech/osism/ceph-ansible:reef\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/ceph-ansible:reef\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: kolla-ansible (Kolla-Ansible Service)\",\n        \"  Expected: registry.osism.tech/osism/kolla-ansible:2024.2\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/kolla-ansible:2024.2\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: osismclient (OSISM Client)\",\n        \"  Expected: registry.osism.tech/osism/osism:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: ara-server (ARA Server)\",\n        \"  Expected: registry.osism.tech/osism/ara-server:1.7.3\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/ara-server:1.7.3\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: mariadb (MariaDB for ARA)\",\n        \"  Expected: registry.osism.tech/dockerhub/library/mariadb:11.8.4\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/dockerhub/library/mariadb:11.8.4\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: frontend (OSISM Frontend)\",\n        \"  Expected: registry.osism.tech/osism/osism-frontend:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism-frontend:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: redis (Redis Cache)\",\n        \"  Expected: registry.osism.tech/dockerhub/library/redis:7.4.7-alpine\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/dockerhub/library/redis:7.4.7-alpine\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: api (OSISM API Service)\",\n        \"  Expected: registry.osism.tech/osism/osism:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: listener (OpenStack Event Listener)\",\n        \"  Expected: registry.osism.tech/osism/osism:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: openstack (OpenStack Integration)\",\n        \"  Expected: registry.osism.tech/osism/osism:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: beat (Celery Beat Scheduler)\",\n        \"  Expected: registry.osism.tech/osism/osism:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: flower (Celery Flower Monitor)\",\n        \"  Expected: registry.osism.tech/osism/osism:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"=== Summary ===\",\n        \"Errors (version mismatches): 0\",\n        \"Warnings (expected containers not running): 0\",\n        \"\",\n        \"\u2705 All running containers match expected versions!\"\n    ]\n}\n\nTASK [osism.services.manager : Skip version check due to service configuration] ***\nskipping: [testbed-manager]\n\nPLAY RECAP *********************************************************************\ntestbed-manager            : ok=70   changed=37   unreachable=0    failed=0    skipped=12   rescued=0    ignored=0   \n\n+ [[ -e /opt/venv/bin/activate ]]\n+ deactivate\n+ '[' -n /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin ']'\n+ PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin\n+ export PATH\n+ unset _OLD_VIRTUAL_PATH\n+ '[' -n '' ']'\n+ hash -r\n+ '[' -n '' ']'\n+ unset VIRTUAL_ENV\n+ unset VIRTUAL_ENV_PROMPT\n+ '[' '!' '' = nondestructive ']'\n+ unset -f deactivate\n+ cp /home/dragon/.ssh/id_rsa.pub /opt/ansible/secrets/id_rsa.operator.pub\n+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]\n+ wait_for_container_healthy 60 ceph-ansible\n+ local max_attempts=60\n+ local name=ceph-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ wait_for_container_healthy 60 kolla-ansible\n+ local max_attempts=60\n+ local name=kolla-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' kolla-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ wait_for_container_healthy 60 osism-ansible\n+ local max_attempts=60\n+ local name=osism-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' osism-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ [[ true == \\t\\r\\u\\e ]]\n+ sh -c /opt/configuration/scripts/disable-ara.sh\n+ docker compose --project-directory /opt/manager ps\nNAME                             IMAGE                                                      COMMAND                  SERVICE                CREATED         STATUS                        PORTS\nceph-ansible                     registry.osism.tech/osism/ceph-ansible:reef                \"/entrypoint.sh osis\u2026\"   ceph-ansible           2 minutes ago   Up About a minute (healthy)   \nkolla-ansible                    registry.osism.tech/osism/kolla-ansible:2024.2             \"/entrypoint.sh osis\u2026\"   kolla-ansible          2 minutes ago   Up About a minute (healthy)   \nmanager-api-1                    registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   api                    2 minutes ago   Up 2 minutes (healthy)        192.168.16.5:8000->8000/tcp\nmanager-ara-server-1             registry.osism.tech/osism/ara-server:1.7.3                 \"sh -c '/wait && /ru\u2026\"   ara-server             2 minutes ago   Up 2 minutes (healthy)        8000/tcp\nmanager-beat-1                   registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   beat                   2 minutes ago   Up 2 minutes (healthy)        \nmanager-flower-1                 registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   flower                 2 minutes ago   Up 2 minutes (healthy)        \nmanager-inventory_reconciler-1   registry.osism.tech/osism/inventory-reconciler:latest      \"/sbin/tini -- /entr\u2026\"   inventory_reconciler   2 minutes ago   Up About a minute (healthy)   \nmanager-listener-1               registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   listener               2 minutes ago   Up 2 minutes (healthy)        \nmanager-mariadb-1                registry.osism.tech/dockerhub/library/mariadb:11.8.4       \"docker-entrypoint.s\u2026\"   mariadb                2 minutes ago   Up 2 minutes (healthy)        3306/tcp\nmanager-openstack-1              registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   openstack              2 minutes ago   Up 2 minutes (healthy)        \nmanager-redis-1                  registry.osism.tech/dockerhub/library/redis:7.4.7-alpine   \"docker-entrypoint.s\u2026\"   redis                  2 minutes ago   Up 2 minutes (healthy)        6379/tcp\nosism-ansible                    registry.osism.tech/osism/osism-ansible:latest             \"/entrypoint.sh osis\u2026\"   osism-ansible          2 minutes ago   Up About a minute (healthy)   \nosism-frontend                   registry.osism.tech/osism/osism-frontend:latest            \"docker-entrypoint.s\u2026\"   frontend               2 minutes ago   Up 2 minutes                  192.168.16.5:3000->3000/tcp\nosism-kubernetes                 registry.osism.tech/osism/osism-kubernetes:latest          \"/entrypoint.sh osis\u2026\"   osism-kubernetes       2 minutes ago   Up About a minute (healthy)   \nosismclient                      registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- sleep\u2026\"   osismclient            2 minutes ago   Up 2 minutes (healthy)        \n++ semver latest 7.0.0\n+ [[ -1 -ge 0 ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ sed -i s/community.general.yaml/osism.commons.still_alive/ /opt/configuration/environments/ansible.cfg\n+ osism apply resolvconf -l testbed-manager\n\u001b[32m2026-03-21 00:20:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of resolvconf.\u001b[0m\n\u001b[32m2026-03-21 00:20:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1cd44190-d184-4b83-bd15-79b0234c9cb3 (resolvconf) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 00:20:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 1cd44190-d184-4b83-bd15-79b0234c9cb3 (resolvconf) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role resolvconf] ***************************************************\n\nTASK [Gathering Facts] *********************************************************\nSaturday 21 March 2026  00:20:36 +0000 (0:00:00.198)       0:00:00.198 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Check minimum and maximum number of name servers] ***\nSaturday 21 March 2026  00:20:41 +0000 (0:00:04.929)       0:00:05.127 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Include resolvconf tasks] *********************\nSaturday 21 March 2026  00:20:41 +0000 (0:00:00.068)       0:00:05.196 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-resolv.yml for testbed-manager\u001b[0m\n\nTASK [osism.commons.resolvconf : Include distribution specific installation tasks] ***\nSaturday 21 March 2026  00:20:41 +0000 (0:00:00.093)       0:00:05.290 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf] ***\nSaturday 21 March 2026  00:20:41 +0000 (0:00:00.079)       0:00:05.370 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Install package systemd-resolved] *************\nSaturday 21 March 2026  00:20:42 +0000 (0:00:01.120)       0:00:06.490 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf] *****\nSaturday 21 March 2026  00:20:43 +0000 (0:00:00.056)       0:00:06.546 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Archive existing file /etc/resolv.conf] *******\nSaturday 21 March 2026  00:20:43 +0000 (0:00:00.509)       0:00:07.056 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf] ***\nSaturday 21 March 2026  00:20:43 +0000 (0:00:00.074)       0:00:07.131 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Copy configuration files] *********************\nSaturday 21 March 2026  00:20:44 +0000 (0:00:00.533)       0:00:07.665 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Start/enable systemd-resolved service] ********\nSaturday 21 March 2026  00:20:45 +0000 (0:00:01.054)       0:00:08.719 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Include distribution specific configuration tasks] ***\nSaturday 21 March 2026  00:20:46 +0000 (0:00:00.920)       0:00:09.640 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.commons.resolvconf : Restart systemd-resolved service] *************\nSaturday 21 March 2026  00:20:46 +0000 (0:00:00.083)       0:00:09.724 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:20:47 +0000 (0:00:01.106)       0:00:10.831 ******** \n=============================================================================== \nGathering Facts --------------------------------------------------------- 4.93s\nosism.commons.resolvconf : Remove packages configuring /etc/resolv.conf --- 1.12s\nosism.commons.resolvconf : Restart systemd-resolved service ------------- 1.11s\nosism.commons.resolvconf : Copy configuration files --------------------- 1.05s\nosism.commons.resolvconf : Start/enable systemd-resolved service -------- 0.92s\nosism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf --- 0.53s\nosism.commons.resolvconf : Retrieve file status of /etc/resolv.conf ----- 0.51s\nosism.commons.resolvconf : Include resolvconf tasks --------------------- 0.09s\nosism.commons.resolvconf : Include distribution specific configuration tasks --- 0.08s\nosism.commons.resolvconf : Include distribution specific installation tasks --- 0.08s\nosism.commons.resolvconf : Archive existing file /etc/resolv.conf ------- 0.07s\nosism.commons.resolvconf : Check minimum and maximum number of name servers --- 0.07s\nosism.commons.resolvconf : Install package systemd-resolved ------------- 0.06s\n+ osism apply sshconfig\n\u001b[32m2026-03-21 00:20:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of sshconfig.\u001b[0m\n\u001b[32m2026-03-21 00:20:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask edf4ac79-9196-4576-ba98-18766b3fa2b1 (sshconfig) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 00:20:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task edf4ac79-9196-4576-ba98-18766b3fa2b1 (sshconfig) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role sshconfig] ****************************************************\n\nTASK [osism.commons.sshconfig : Get home directory of operator user] ***********\nSaturday 21 March 2026  00:21:01 +0000 (0:00:00.144)       0:00:00.144 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.sshconfig : Ensure .ssh/config.d exist] ********************\nSaturday 21 March 2026  00:21:02 +0000 (0:00:00.859)       0:00:01.004 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.sshconfig : Ensure config for each host exist] *************\nSaturday 21 March 2026  00:21:03 +0000 (0:00:00.474)       0:00:01.479 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5)\u001b[0m\n\nTASK [osism.commons.sshconfig : Add extra config] ******************************\nSaturday 21 March 2026  00:21:08 +0000 (0:00:05.164)       0:00:06.643 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.sshconfig : Assemble ssh config] ***************************\nSaturday 21 March 2026  00:21:08 +0000 (0:00:00.108)       0:00:06.752 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=4   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:21:09 +0000 (0:00:00.516)       0:00:07.268 ******** \n=============================================================================== \nosism.commons.sshconfig : Ensure config for each host exist ------------- 5.16s\nosism.commons.sshconfig : Get home directory of operator user ----------- 0.86s\nosism.commons.sshconfig : Assemble ssh config --------------------------- 0.52s\nosism.commons.sshconfig : Ensure .ssh/config.d exist -------------------- 0.47s\nosism.commons.sshconfig : Add extra config ------------------------------ 0.11s\n+ osism apply known-hosts\n\u001b[32m2026-03-21 00:21:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of known-hosts.\u001b[0m\n\u001b[32m2026-03-21 00:21:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c84cdb3b-fb8b-4d97-9693-70eeb7a327d5 (known-hosts) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 00:21:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task c84cdb3b-fb8b-4d97-9693-70eeb7a327d5 (known-hosts) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role known_hosts] **************************************************\n\nTASK [osism.commons.known_hosts : Run ssh-keyscan for all hosts with hostname] ***\nSaturday 21 March 2026  00:21:23 +0000 (0:00:00.193)       0:00:00.193 ******** \n\u001b[0;32mok: [testbed-manager] => (item=testbed-manager)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-5)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with hostname] ***\nSaturday 21 March 2026  00:21:30 +0000 (0:00:06.436)       0:00:06.629 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-manager)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-0)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-1)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-2)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-3)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-4)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-5)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nSaturday 21 March 2026  00:21:30 +0000 (0:00:00.192)       0:00:06.821 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBGCW8S4XhFwtxakQShgXgO7SwC1ojlJgYP5uttdiBrf1czDr4b2JfGHDaBJkhAR0HWZVp4aSpl8XO7vpRuLHcKg=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCRlOzWKiVjhf3/9gUxrgBZmRlENj/bm0EeQnimGnkItBo13vP/sIQySfQ4wuvVN2E7kjt0LDhutlgBP+9+9QmN7QyxACxIVy+Yv3TDwZaTwkeU71pgeZ2nuphSGxEwWxtpOk6NTrdk8LCUuXK5IzKa++u+/ESIaPuIjMUQpLsc7m6taSCDlQwhMPbh4pkUCOfTKvWTZvUTzhfXd0v2yu2oYE6HKi552UF1xZeFUdWj2mLvuFVtObaVPiqDrfl80u0mOYPZzkeHGg4OwJqr8+eYNrTy2PPcvpPJz2zd6KCKrScr0YqlFrI1/iqMEdB24un2H9hiGm21ETl7jBPeYUSkY8fwR/z18SGIEqcwxLwdyEVufwpclNnJpLdeueDqOTUWSzm/2eV9xxQRUHv37+cdQHGlsw9Vkd1vB7S+RyiMCxHvzM+F7s9qheuk1oN1YBk0Gu9PirJddsCb13/vR9Uq3aGR8Rl+JNQjBLZ52WoiBOU6x9Xwl06LIMXvTYbk6R8=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIINhMRJZFRnFvt0mw99z3/Oe5sZY19PiSKPeN08h+FZs)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nSaturday 21 March 2026  00:21:31 +0000 (0:00:01.224)       0:00:08.046 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDBFYMv4DFwxq5kkUW4Yk1dzqvK6rxV/EnP+KlAQYoqgq0Jxy6RiZBjZYrs44Cz5N22qgQKRJOIKZeviB2HMPXP9M/stCMvrwKHQ759UIXdlCFSM1wb/XiWVEfpvyg17SK8C6Sk3TFQNxDBvH8tuoB9itW69HmiDzMNQzJjw+8rbaAFfrJVxweUkVfjVL3u1XXMaTcarhWQ8AM1qCt8WSpc87HZ+5nPz6MiwzvnlbXpGvmPDH8jpVGGI0liY0uuXbc+cVSyVHpHqcS+0btiGc2o8CrV86oBz1kBA2QlztckLU8o0J9EsO9tXNBFf9dXpgLoq/3hsyFh5J3zWQK6FlyaRX0fiwTKLyPuoeaIn8rjM6YLfZCdJW+dFfo7mmN4FJaLcQz0flwy5rGNztyjK76Zij71zy4a5YNJoqnoHlOkp12AwJZVX+UrXTpPircM/LObMS26OBZousjzYSsowqdR1qwHJkth5DI9wqPdIj0OYb8w422VthxdB7jWvJDL+yM=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBHh3FDIOKYfKw5RWzPvfahw1uwX3EnFPFc7j151CAfmuNXNhWNd/g7+a+BcFHxWoUVLQ2bHNkBvDQ853xelpFss=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIJtBRNCPeE2jW0fyck0ZOiM7FLqI8afu36mE896gdALV)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nSaturday 21 March 2026  00:21:32 +0000 (0:00:01.021)       0:00:09.068 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCkOmCQltaAtHo3qnWk8Zva2/YySzFSuTk33gJ8qFQSOoJ/JMivrQOeeT/xhyTpT0KdDa4wUEJ2DFsRRJtQ4RJT1tOvvKVcOHCUAb5TukWWcRLzojdpJje3FzjBhm3CIYeeCVmUFQob4+Fbsvq14fYAJwJ+iLRVb3WYjJ2DQpdgnSizr+YXC1FgaUAHsI0qKtZXRFHs3kRz/V086+0QhcVvLEVkzCcrUJpK3br38IxBNtUt4riGvSJeBAEZjPDCoZYULNKhfuDJpSTXtFhDPURIdVGMiphFAyJc75UXDmAur9PgARiKpIUiW6lkK99pEzQL81mDPGgfR2IVt6WyHB6+KHsyGBkAhV9Hl/mu0wPs/BQJIiwzKKEZwjwQHZtH2WbPY5ZO01+mbR2B/mdJIyIBPOT70tA3oF/i1dOm2akpHMA5b6Ygn6LgNXZ2/R/ONzk5FqfWdjAyR/g8trdF5OeFSbs8+yZjmixmGFlXUgzP09SPhzcaZr6eu6Re8ubRUCk=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPidv0r0leu7YKpNvc3qywT5qo1UxOebZ4/1THmXIT8p95UG8xxtEgxjB6LiDlrIKHnOQcrjX0aPZd9tK0OUMH0=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIBXGRQ9OqD5v29iwd1ZjIplLtCmKlBxfmQ/Z4YtqKcvG)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nSaturday 21 March 2026  00:21:33 +0000 (0:00:01.046)       0:00:10.114 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDLUD36gZZ07RR35hp8bzvE4EMQSFjQPExYfzrUx2UDw1Iw7hZO7DIVqdOYeO71sHdJCALdChBOd+sqrPhZVHU2pte0NRFLiKlfHPGtT8En+gvydegKX9ZrbDxH/QDQzAoK3oxa6rSY/hgrjalxSfFsoX3ck0+B+BAsnPQkl7pJb3qMxKbwj0BZg8hvl3ORoU2LywtrmZCwnldNdexRDw9WuG4A7CUtBAw0sRk1F+6XMVDBpgPy7Z/VKL/77GNdzZhfHnWRoZR16U6OiM4km40O9HhGpfguJXF7QS5O+I1qOeLnqkEY8C0Iw6VfLDril4+55vdrRO/301+/2sR34VdUinHwzvz8CK/gfAQFfOHQKtD5DbfNdAYJjchZmSd5QsQcJNCk+Oi3zJ+W30kkTnRWfK+X7O4YZ1Vr14gfuhEpWImGUVEKZ1MqIcf6qoCKGlAbz4Ftv6g9CAy+X2LQgMrYK7AsSPDFwQ2IXLkpNB8885jAXlajrYNFi+BuHilCJjM=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBKVwUgL19RWciqX6jXMbqinVkklJq2z1AvXg256NiASkiG1Mj5ak7uFZEkNGaEHf+K1QBoUh7fbzGpgkRUarIPU=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDo385EdO3Q9a9gM7vB54U7tO40OJjhIKuCD/4Lczktc)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nSaturday 21 March 2026  00:21:34 +0000 (0:00:01.037)       0:00:11.151 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBMw16gD3Pbm/obmA1y0WkDRYOE6nz7IznqcYxNJs1Ttz0TfWx4+l2TZZB+cWEZlqH1wXHeH4gIujSHjYb0z47x4=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCzB2yNwrdxiybJKIFTUFKw9hgA3eTgnpnh7Bnn2LQx2kmV9Tb3SUpq4hu3HGINP8vFI02QBpLJY3J8calXDdlmDA7s+90KFx+ituDeWizWVTuBtnDfWKH+bXqnpK4fytWz+rraBQg0zENXnkq95/sFYQAlEfklsEN6QtaTqx9UBFXWyST4nveGW9WLbfvqaBiGw67W1alnRV39FUhctaQRVoxfGLmERo7vX7PAhq+olypDXKriYYrJUJYej+TBdEhBQKbjHxl+ltGYYsCAgQ3YNMicYsutcKJq1fn5y1xhkGh2D/3jMSj8zE3H/ZbwyEuwMvRce6WxPHyfRwwEEcw1z+ZWS2fr565/aK/d+oa+dPgmHdAkOHTLTTw5GHE32YXuhZ+oqUORH5NyTeTtl0jj+JFCUt1rEN+4ZyNQQtIX4k4bxce/K7hFeAmBE3oxWq8RaaR95g89U5m1CmJ22hK1sCZ6+4kHW4gmhwZU0tVuv+dGZHtzxNdHQhnjrFnkEfk=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIPChxonu3Hcdz1NXmezrajxuZTmJk01UE7s701krwJ8B)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nSaturday 21 March 2026  00:21:35 +0000 (0:00:01.037)       0:00:12.189 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLaFmk9jZnhiVtKL0wRcGE4qVsljEsur5kTWwGDIJ1BGW9608838oko7fKNiNcyELZFzn+EJxiyJc06R/YAnm94=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAILcvUHyiPwzNQOC7IOncTN/c7TkgoFEUaxcoF1QveIN9)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC+2o4j0grHvLW3HhVtn4eHEakT9YYI3Sb++Ci0amagE5qNEbs/ZxyR9lzC65lYaEb3k10xr1I7dsrCB4xGyg89vnvU9BhFTChCpOBN4XSsQI0VXJcoK1gwh1vVcAgzoYbDh9qj25CpXEKmPS79tu5aMhzf7ENIjYP84/DyI7o/3zRk3GIocva42Ojjj8/EmmhNU0q+tfq4+kHnI6xCSoDK9YINST/sakm5VZWgzvZIluw7dbpbOvCgfflUmbaoO2emen9a1T23P/2RQLaCnVoiOij5+TIM4yZCPtFmg8Pf+BgRPt7dkjngX62KzF5NbXSNQGD7jI7L2bBWaOdKx1SmxbjIbQ/+ZwOcaUtfC49XfHHRrCQv6WXziDCPOSe+11xrCaHKcJMEztUEjJAR1Q0PI/bShfT+8TsULIMm1PYdKL+97cN57ccRrW2BEh1UExjG1LpvT+F0ihjouDQtOJdQTvGJHfAIwk0ttixM13ny9r/YvhZcZ8+OHpx75ljTc20=)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nSaturday 21 March 2026  00:21:36 +0000 (0:00:01.039)       0:00:13.228 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBJcLJ530VGI+fZqYZivjtRjY60XsOdACkcMGZCO6DM1QCtAMis3Vt7RkgrBST8pYU1j4GVKr/b830ZCxBx8OWD8=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC8TUaJpxO/zwNXdrLifgopr+lsK/T9T7LNe2Pa4nLlZD69HGd8aTTUK8isoyIT+NnnT1Hna9SMLxNv8ibVyw6P7m3g8I6jzlc+J5ex443F1eIuwKlHiLKGrCCX+zp7DFbbJhcXDfl8cglHXPgdeyGVbDEWqOmWRSalh7QWAdF4dhVKp2J+rmuqJEz3MxqjNfpelItlm8Cd9dAL22kjI8YeqXTWW/q/5bkEwvApD5PqoxOe8FpSPtQUW1xx9/pwWyIbsN6ARcX8nrmzAHbwpHvc0GR39x1ahcLnFm+7UE+5Mxwq8SvCTQe2oeLkEg7XRdj3nJ7cUmGFTGczVMHULU6ZuHiabMhbwVbBBnjm5mfkCcOVdMOaYEAcqUBB51w/6RQpB+I23Di40MibLigFEoQOaIEHZtggAivOaXouqPlNQ21MevgBw1jzwVNtR44U4uBS2uMSrCeopIuxVYptnlOFa3YdDeS/gfAacyD3/dkLymC0rC4QBn1PjV46J6bbQ20=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIIdSvMWwXHWpnHDy//kDiEa/cHDwBq24RvQjo19fNzjB)\u001b[0m\n\nTASK [osism.commons.known_hosts : Run ssh-keyscan for all hosts with ansible_host] ***\nSaturday 21 March 2026  00:21:38 +0000 (0:00:01.072)       0:00:14.300 ******** \n\u001b[0;32mok: [testbed-manager] => (item=testbed-manager)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-5)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with ansible_host] ***\nSaturday 21 March 2026  00:21:43 +0000 (0:00:05.317)       0:00:19.618 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-manager)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-0)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-1)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-2)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-3)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-4)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-5)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nSaturday 21 March 2026  00:21:43 +0000 (0:00:00.169)       0:00:19.787 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCRlOzWKiVjhf3/9gUxrgBZmRlENj/bm0EeQnimGnkItBo13vP/sIQySfQ4wuvVN2E7kjt0LDhutlgBP+9+9QmN7QyxACxIVy+Yv3TDwZaTwkeU71pgeZ2nuphSGxEwWxtpOk6NTrdk8LCUuXK5IzKa++u+/ESIaPuIjMUQpLsc7m6taSCDlQwhMPbh4pkUCOfTKvWTZvUTzhfXd0v2yu2oYE6HKi552UF1xZeFUdWj2mLvuFVtObaVPiqDrfl80u0mOYPZzkeHGg4OwJqr8+eYNrTy2PPcvpPJz2zd6KCKrScr0YqlFrI1/iqMEdB24un2H9hiGm21ETl7jBPeYUSkY8fwR/z18SGIEqcwxLwdyEVufwpclNnJpLdeueDqOTUWSzm/2eV9xxQRUHv37+cdQHGlsw9Vkd1vB7S+RyiMCxHvzM+F7s9qheuk1oN1YBk0Gu9PirJddsCb13/vR9Uq3aGR8Rl+JNQjBLZ52WoiBOU6x9Xwl06LIMXvTYbk6R8=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBGCW8S4XhFwtxakQShgXgO7SwC1ojlJgYP5uttdiBrf1czDr4b2JfGHDaBJkhAR0HWZVp4aSpl8XO7vpRuLHcKg=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIINhMRJZFRnFvt0mw99z3/Oe5sZY19PiSKPeN08h+FZs)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nSaturday 21 March 2026  00:21:44 +0000 (0:00:01.025)       0:00:20.812 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDBFYMv4DFwxq5kkUW4Yk1dzqvK6rxV/EnP+KlAQYoqgq0Jxy6RiZBjZYrs44Cz5N22qgQKRJOIKZeviB2HMPXP9M/stCMvrwKHQ759UIXdlCFSM1wb/XiWVEfpvyg17SK8C6Sk3TFQNxDBvH8tuoB9itW69HmiDzMNQzJjw+8rbaAFfrJVxweUkVfjVL3u1XXMaTcarhWQ8AM1qCt8WSpc87HZ+5nPz6MiwzvnlbXpGvmPDH8jpVGGI0liY0uuXbc+cVSyVHpHqcS+0btiGc2o8CrV86oBz1kBA2QlztckLU8o0J9EsO9tXNBFf9dXpgLoq/3hsyFh5J3zWQK6FlyaRX0fiwTKLyPuoeaIn8rjM6YLfZCdJW+dFfo7mmN4FJaLcQz0flwy5rGNztyjK76Zij71zy4a5YNJoqnoHlOkp12AwJZVX+UrXTpPircM/LObMS26OBZousjzYSsowqdR1qwHJkth5DI9wqPdIj0OYb8w422VthxdB7jWvJDL+yM=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBHh3FDIOKYfKw5RWzPvfahw1uwX3EnFPFc7j151CAfmuNXNhWNd/g7+a+BcFHxWoUVLQ2bHNkBvDQ853xelpFss=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIJtBRNCPeE2jW0fyck0ZOiM7FLqI8afu36mE896gdALV)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nSaturday 21 March 2026  00:21:45 +0000 (0:00:01.034)       0:00:21.847 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIBXGRQ9OqD5v29iwd1ZjIplLtCmKlBxfmQ/Z4YtqKcvG)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCkOmCQltaAtHo3qnWk8Zva2/YySzFSuTk33gJ8qFQSOoJ/JMivrQOeeT/xhyTpT0KdDa4wUEJ2DFsRRJtQ4RJT1tOvvKVcOHCUAb5TukWWcRLzojdpJje3FzjBhm3CIYeeCVmUFQob4+Fbsvq14fYAJwJ+iLRVb3WYjJ2DQpdgnSizr+YXC1FgaUAHsI0qKtZXRFHs3kRz/V086+0QhcVvLEVkzCcrUJpK3br38IxBNtUt4riGvSJeBAEZjPDCoZYULNKhfuDJpSTXtFhDPURIdVGMiphFAyJc75UXDmAur9PgARiKpIUiW6lkK99pEzQL81mDPGgfR2IVt6WyHB6+KHsyGBkAhV9Hl/mu0wPs/BQJIiwzKKEZwjwQHZtH2WbPY5ZO01+mbR2B/mdJIyIBPOT70tA3oF/i1dOm2akpHMA5b6Ygn6LgNXZ2/R/ONzk5FqfWdjAyR/g8trdF5OeFSbs8+yZjmixmGFlXUgzP09SPhzcaZr6eu6Re8ubRUCk=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPidv0r0leu7YKpNvc3qywT5qo1UxOebZ4/1THmXIT8p95UG8xxtEgxjB6LiDlrIKHnOQcrjX0aPZd9tK0OUMH0=)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nSaturday 21 March 2026  00:21:46 +0000 (0:00:01.021)       0:00:22.869 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBKVwUgL19RWciqX6jXMbqinVkklJq2z1AvXg256NiASkiG1Mj5ak7uFZEkNGaEHf+K1QBoUh7fbzGpgkRUarIPU=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDLUD36gZZ07RR35hp8bzvE4EMQSFjQPExYfzrUx2UDw1Iw7hZO7DIVqdOYeO71sHdJCALdChBOd+sqrPhZVHU2pte0NRFLiKlfHPGtT8En+gvydegKX9ZrbDxH/QDQzAoK3oxa6rSY/hgrjalxSfFsoX3ck0+B+BAsnPQkl7pJb3qMxKbwj0BZg8hvl3ORoU2LywtrmZCwnldNdexRDw9WuG4A7CUtBAw0sRk1F+6XMVDBpgPy7Z/VKL/77GNdzZhfHnWRoZR16U6OiM4km40O9HhGpfguJXF7QS5O+I1qOeLnqkEY8C0Iw6VfLDril4+55vdrRO/301+/2sR34VdUinHwzvz8CK/gfAQFfOHQKtD5DbfNdAYJjchZmSd5QsQcJNCk+Oi3zJ+W30kkTnRWfK+X7O4YZ1Vr14gfuhEpWImGUVEKZ1MqIcf6qoCKGlAbz4Ftv6g9CAy+X2LQgMrYK7AsSPDFwQ2IXLkpNB8885jAXlajrYNFi+BuHilCJjM=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDo385EdO3Q9a9gM7vB54U7tO40OJjhIKuCD/4Lczktc)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nSaturday 21 March 2026  00:21:47 +0000 (0:00:01.013)       0:00:23.882 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCzB2yNwrdxiybJKIFTUFKw9hgA3eTgnpnh7Bnn2LQx2kmV9Tb3SUpq4hu3HGINP8vFI02QBpLJY3J8calXDdlmDA7s+90KFx+ituDeWizWVTuBtnDfWKH+bXqnpK4fytWz+rraBQg0zENXnkq95/sFYQAlEfklsEN6QtaTqx9UBFXWyST4nveGW9WLbfvqaBiGw67W1alnRV39FUhctaQRVoxfGLmERo7vX7PAhq+olypDXKriYYrJUJYej+TBdEhBQKbjHxl+ltGYYsCAgQ3YNMicYsutcKJq1fn5y1xhkGh2D/3jMSj8zE3H/ZbwyEuwMvRce6WxPHyfRwwEEcw1z+ZWS2fr565/aK/d+oa+dPgmHdAkOHTLTTw5GHE32YXuhZ+oqUORH5NyTeTtl0jj+JFCUt1rEN+4ZyNQQtIX4k4bxce/K7hFeAmBE3oxWq8RaaR95g89U5m1CmJ22hK1sCZ6+4kHW4gmhwZU0tVuv+dGZHtzxNdHQhnjrFnkEfk=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBMw16gD3Pbm/obmA1y0WkDRYOE6nz7IznqcYxNJs1Ttz0TfWx4+l2TZZB+cWEZlqH1wXHeH4gIujSHjYb0z47x4=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIPChxonu3Hcdz1NXmezrajxuZTmJk01UE7s701krwJ8B)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nSaturday 21 March 2026  00:21:48 +0000 (0:00:01.032)       0:00:24.915 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAILcvUHyiPwzNQOC7IOncTN/c7TkgoFEUaxcoF1QveIN9)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC+2o4j0grHvLW3HhVtn4eHEakT9YYI3Sb++Ci0amagE5qNEbs/ZxyR9lzC65lYaEb3k10xr1I7dsrCB4xGyg89vnvU9BhFTChCpOBN4XSsQI0VXJcoK1gwh1vVcAgzoYbDh9qj25CpXEKmPS79tu5aMhzf7ENIjYP84/DyI7o/3zRk3GIocva42Ojjj8/EmmhNU0q+tfq4+kHnI6xCSoDK9YINST/sakm5VZWgzvZIluw7dbpbOvCgfflUmbaoO2emen9a1T23P/2RQLaCnVoiOij5+TIM4yZCPtFmg8Pf+BgRPt7dkjngX62KzF5NbXSNQGD7jI7L2bBWaOdKx1SmxbjIbQ/+ZwOcaUtfC49XfHHRrCQv6WXziDCPOSe+11xrCaHKcJMEztUEjJAR1Q0PI/bShfT+8TsULIMm1PYdKL+97cN57ccRrW2BEh1UExjG1LpvT+F0ihjouDQtOJdQTvGJHfAIwk0ttixM13ny9r/YvhZcZ8+OHpx75ljTc20=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLaFmk9jZnhiVtKL0wRcGE4qVsljEsur5kTWwGDIJ1BGW9608838oko7fKNiNcyELZFzn+EJxiyJc06R/YAnm94=)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nSaturday 21 March 2026  00:21:49 +0000 (0:00:01.043)       0:00:25.958 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC8TUaJpxO/zwNXdrLifgopr+lsK/T9T7LNe2Pa4nLlZD69HGd8aTTUK8isoyIT+NnnT1Hna9SMLxNv8ibVyw6P7m3g8I6jzlc+J5ex443F1eIuwKlHiLKGrCCX+zp7DFbbJhcXDfl8cglHXPgdeyGVbDEWqOmWRSalh7QWAdF4dhVKp2J+rmuqJEz3MxqjNfpelItlm8Cd9dAL22kjI8YeqXTWW/q/5bkEwvApD5PqoxOe8FpSPtQUW1xx9/pwWyIbsN6ARcX8nrmzAHbwpHvc0GR39x1ahcLnFm+7UE+5Mxwq8SvCTQe2oeLkEg7XRdj3nJ7cUmGFTGczVMHULU6ZuHiabMhbwVbBBnjm5mfkCcOVdMOaYEAcqUBB51w/6RQpB+I23Di40MibLigFEoQOaIEHZtggAivOaXouqPlNQ21MevgBw1jzwVNtR44U4uBS2uMSrCeopIuxVYptnlOFa3YdDeS/gfAacyD3/dkLymC0rC4QBn1PjV46J6bbQ20=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBJcLJ530VGI+fZqYZivjtRjY60XsOdACkcMGZCO6DM1QCtAMis3Vt7RkgrBST8pYU1j4GVKr/b830ZCxBx8OWD8=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIIdSvMWwXHWpnHDy//kDiEa/cHDwBq24RvQjo19fNzjB)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write static known_hosts entries] ************\nSaturday 21 March 2026  00:21:50 +0000 (0:00:01.043)       0:00:27.002 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.known_hosts : Write extra known_hosts entries] *************\nSaturday 21 March 2026  00:21:50 +0000 (0:00:00.172)       0:00:27.175 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.known_hosts : Delete known_hosts entries] ******************\nSaturday 21 March 2026  00:21:50 +0000 (0:00:00.065)       0:00:27.240 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.known_hosts : Set file permissions] ************************\nSaturday 21 March 2026  00:21:51 +0000 (0:00:00.056)       0:00:27.296 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=31  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:21:51 +0000 (0:00:00.512)       0:00:27.809 ******** \n=============================================================================== \nosism.commons.known_hosts : Run ssh-keyscan for all hosts with hostname --- 6.44s\nosism.commons.known_hosts : Run ssh-keyscan for all hosts with ansible_host --- 5.32s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.22s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.07s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.05s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.04s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.04s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.04s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.04s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.04s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.04s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.03s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.03s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.02s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.02s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.01s\nosism.commons.known_hosts : Set file permissions ------------------------ 0.51s\nosism.commons.known_hosts : Write scanned known_hosts entries for all hosts with hostname --- 0.19s\nosism.commons.known_hosts : Write static known_hosts entries ------------ 0.17s\nosism.commons.known_hosts : Write scanned known_hosts entries for all hosts with ansible_host --- 0.17s\n+ osism apply squid\n\u001b[32m2026-03-21 00:22:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of squid.\u001b[0m\n\u001b[32m2026-03-21 00:22:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d030f99f-1670-411b-a7f7-ce57e9141b9b (squid) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 00:22:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task d030f99f-1670-411b-a7f7-ce57e9141b9b (squid) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role squid] ********************************************************\n\nTASK [osism.services.squid : Include install tasks] ****************************\nSaturday 21 March 2026  00:22:06 +0000 (0:00:00.187)       0:00:00.187 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/squid/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.services.squid : Install required packages] ************************\nSaturday 21 March 2026  00:22:06 +0000 (0:00:00.079)       0:00:00.267 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.squid : Create required directories] **********************\nSaturday 21 March 2026  00:22:08 +0000 (0:00:02.335)       0:00:02.603 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/squid/configuration)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/squid/configuration/conf.d)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/squid)\u001b[0m\n\nTASK [osism.services.squid : Copy squid configuration files] *******************\nSaturday 21 March 2026  00:22:09 +0000 (0:00:01.168)       0:00:03.771 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=osism.conf)\u001b[0m\n\nTASK [osism.services.squid : Remove osism_allow_list.conf configuration file] ***\nSaturday 21 March 2026  00:22:10 +0000 (0:00:01.027)       0:00:04.798 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.squid : Copy docker-compose.yml file] *********************\nSaturday 21 March 2026  00:22:11 +0000 (0:00:00.341)       0:00:05.140 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.squid : Manage squid service] *****************************\nSaturday 21 March 2026  00:22:12 +0000 (0:00:00.884)       0:00:06.024 ******** \n\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage squid service (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.squid : Restart squid service] *****************\nSaturday 21 March 2026  00:22:43 +0000 (0:00:30.911)       0:00:36.935 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.squid : Wait for squid service to start] *******\nSaturday 21 March 2026  00:22:58 +0000 (0:00:15.784)       0:00:52.719 ******** \nPausing for 60 seconds\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.squid : Register that squid service was restarted] ***\nSaturday 21 March 2026  00:23:58 +0000 (0:01:00.084)       0:01:52.804 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.squid : Wait for an healthy squid service] *****\nSaturday 21 March 2026  00:23:59 +0000 (0:00:00.060)       0:01:52.864 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=11  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:23:59 +0000 (0:00:00.585)       0:01:53.450 ******** \n=============================================================================== \nosism.services.squid : Wait for squid service to start ----------------- 60.08s\nosism.services.squid : Manage squid service ---------------------------- 30.91s\nosism.services.squid : Restart squid service --------------------------- 15.78s\nosism.services.squid : Install required packages ------------------------ 2.34s\nosism.services.squid : Create required directories ---------------------- 1.17s\nosism.services.squid : Copy squid configuration files ------------------- 1.03s\nosism.services.squid : Copy docker-compose.yml file --------------------- 0.88s\nosism.services.squid : Wait for an healthy squid service ---------------- 0.59s\nosism.services.squid : Remove osism_allow_list.conf configuration file --- 0.34s\nosism.services.squid : Include install tasks ---------------------------- 0.08s\nosism.services.squid : Register that squid service was restarted -------- 0.06s\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ /opt/configuration/scripts/set-kolla-namespace.sh kolla\n+ set -e\n+ NAMESPACE=kolla\n+ sed -i 's#docker_namespace: .*#docker_namespace: kolla#g' /opt/configuration/inventory/group_vars/all/kolla.yml\n++ semver latest 9.0.0\n+ [[ -1 -lt 0 ]]\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ osism apply operator -u ubuntu -l testbed-nodes\n\u001b[32m2026-03-21 00:24:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of operator.\u001b[0m\n\u001b[32m2026-03-21 00:24:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ec44afcf-bc56-470f-92a0-18c80be44f09 (operator) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 00:24:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task ec44afcf-bc56-470f-92a0-18c80be44f09 (operator) has been started and output is visible here.\u001b[0m\n\nPLAY [Make ssh pipelining working] *********************************************\n\nTASK [Gathering Facts] *********************************************************\nSaturday 21 March 2026  00:24:14 +0000 (0:00:00.169)       0:00:00.169 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Do not require tty for all users] ****************************************\nSaturday 21 March 2026  00:24:17 +0000 (0:00:03.447)       0:00:03.617 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role operator] *****************************************************\n\nTASK [osism.commons.operator : Gather variables for each operating system] *****\nSaturday 21 March 2026  00:24:18 +0000 (0:00:00.916)       0:00:04.533 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set operator_groups variable to default value] ***\nSaturday 21 March 2026  00:24:18 +0000 (0:00:00.157)       0:00:04.691 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Create operator group] **************************\nSaturday 21 March 2026  00:24:19 +0000 (0:00:00.137)       0:00:04.828 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.operator : Create user] ************************************\nSaturday 21 March 2026  00:24:19 +0000 (0:00:00.702)       0:00:05.530 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.operator : Add user to additional groups] ******************\nSaturday 21 March 2026  00:24:20 +0000 (0:00:00.995)       0:00:06.526 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=sudo)\u001b[0m\n\nTASK [osism.commons.operator : Copy user sudoers file] *************************\nSaturday 21 March 2026  00:24:22 +0000 (0:00:01.278)       0:00:07.805 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.operator : Set language variables in .bashrc configuration file] ***\nSaturday 21 March 2026  00:24:23 +0000 (0:00:01.335)       0:00:09.140 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[1;35m[WARNING]: Module remote_tmp /root/.ansible/tmp did not exist and was created\u001b[0m\n\u001b[1;35mwith a mode of 0700, this may cause issues when running as another user. To\u001b[0m\n\u001b[1;35mavoid this, create the remote_tmp dir with the correct permissions manually\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\nTASK [osism.commons.operator : Set custom environment variables in .bashrc configuration file] ***\nSaturday 21 March 2026  00:24:24 +0000 (0:00:01.251)       0:00:10.392 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file] ***\nSaturday 21 March 2026  00:24:24 +0000 (0:00:00.144)       0:00:10.537 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Create .ssh directory] **************************\nSaturday 21 March 2026  00:24:24 +0000 (0:00:00.151)       0:00:10.688 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Check number of SSH authorized keys] ************\nSaturday 21 March 2026  00:24:25 +0000 (0:00:00.529)       0:00:11.218 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set ssh authorized keys] ************************\nSaturday 21 March 2026  00:24:25 +0000 (0:00:00.183)       0:00:11.401 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Delete ssh authorized keys] *********************\nSaturday 21 March 2026  00:24:26 +0000 (0:00:00.716)       0:00:12.118 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set authorized GitHub accounts] *****************\nSaturday 21 March 2026  00:24:26 +0000 (0:00:00.144)       0:00:12.263 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Delete authorized GitHub accounts] **************\nSaturday 21 March 2026  00:24:26 +0000 (0:00:00.144)       0:00:12.407 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set password] ***********************************\nSaturday 21 March 2026  00:24:26 +0000 (0:00:00.133)       0:00:12.541 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.operator : Unset & lock password] **************************\nSaturday 21 March 2026  00:24:27 +0000 (0:00:00.653)       0:00:13.194 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:24:27 +0000 (0:00:00.235)       0:00:13.429 ******** \n=============================================================================== \nGathering Facts --------------------------------------------------------- 3.45s\nosism.commons.operator : Copy user sudoers file ------------------------- 1.34s\nosism.commons.operator : Add user to additional groups ------------------ 1.28s\nosism.commons.operator : Set language variables in .bashrc configuration file --- 1.25s\nosism.commons.operator : Create user ------------------------------------ 1.00s\nDo not require tty for all users ---------------------------------------- 0.92s\nosism.commons.operator : Set ssh authorized keys ------------------------ 0.72s\nosism.commons.operator : Create operator group -------------------------- 0.70s\nosism.commons.operator : Set password ----------------------------------- 0.65s\nosism.commons.operator : Create .ssh directory -------------------------- 0.53s\nosism.commons.operator : Unset & lock password -------------------------- 0.24s\nosism.commons.operator : Check number of SSH authorized keys ------------ 0.18s\nosism.commons.operator : Gather variables for each operating system ----- 0.16s\nosism.commons.operator : Set custom PS1 prompt in .bashrc configuration file --- 0.15s\nosism.commons.operator : Delete ssh authorized keys --------------------- 0.14s\nosism.commons.operator : Set custom environment variables in .bashrc configuration file --- 0.14s\nosism.commons.operator : Set authorized GitHub accounts ----------------- 0.14s\nosism.commons.operator : Set operator_groups variable to default value --- 0.14s\nosism.commons.operator : Delete authorized GitHub accounts -------------- 0.13s\n+ osism apply --environment custom facts\n\u001b[32m2026-03-21 00:24:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play facts in environment custom\u001b[0m\n\u001b[32m2026-03-21 00:24:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of facts.\u001b[0m\n\u001b[32m2026-03-21 00:24:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9c35476d-6400-4b54-acfb-b274a3f36f36 (facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 00:24:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 9c35476d-6400-4b54-acfb-b274a3f36f36 (facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Copy custom network devices fact] ****************************************\n\nTASK [Create custom facts directory] *******************************************\nSaturday 21 March 2026  00:24:42 +0000 (0:00:00.118)       0:00:00.118 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [Copy fact file] **********************************************************\nSaturday 21 March 2026  00:24:43 +0000 (0:00:01.350)       0:00:01.469 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nPLAY [Copy custom ceph devices facts] ******************************************\n\nTASK [osism.commons.repository : Gather variables for each operating system] ***\nSaturday 21 March 2026  00:24:45 +0000 (0:00:01.345)       0:00:02.815 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Set repository_default fact to default value] ***\nSaturday 21 March 2026  00:24:45 +0000 (0:00:00.096)       0:00:02.911 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Set repositories to default] ******************\nSaturday 21 March 2026  00:24:45 +0000 (0:00:00.195)       0:00:03.106 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Include distribution specific repository tasks] ***\nSaturday 21 March 2026  00:24:45 +0000 (0:00:00.194)       0:00:03.301 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****\nSaturday 21 March 2026  00:24:45 +0000 (0:00:00.117)       0:00:03.419 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************\nSaturday 21 March 2026  00:24:46 +0000 (0:00:00.425)       0:00:03.844 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Copy 99osism apt configuration] ***************\nSaturday 21 March 2026  00:24:46 +0000 (0:00:00.115)       0:00:03.960 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Remove sources.list file] *********************\nSaturday 21 March 2026  00:24:47 +0000 (0:00:01.062)       0:00:05.022 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Copy ubuntu.sources file] *********************\nSaturday 21 March 2026  00:24:48 +0000 (0:00:00.496)       0:00:05.519 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Update package cache] *************************\nSaturday 21 March 2026  00:24:49 +0000 (0:00:01.139)       0:00:06.659 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Install required packages (RedHat)] **************************************\nSaturday 21 March 2026  00:25:06 +0000 (0:00:17.434)       0:00:24.093 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Install required packages (Debian)] **************************************\nSaturday 21 March 2026  00:25:06 +0000 (0:00:00.091)       0:00:24.184 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [Create custom facts directory] *******************************************\nSaturday 21 March 2026  00:25:13 +0000 (0:00:07.285)       0:00:31.469 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Copy fact files] *********************************************************\nSaturday 21 March 2026  00:25:14 +0000 (0:00:00.487)       0:00:31.957 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_osd_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_osd_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_osd_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_osd_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_osd_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_osd_devices_all)\u001b[0m\n\nRUNNING HANDLER [osism.commons.repository : Force update of package cache] *****\nSaturday 21 March 2026  00:25:18 +0000 (0:00:03.802)       0:00:35.759 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nSaturday 21 March 2026  00:25:19 +0000 (0:00:01.551)       0:00:37.311 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:25:23 +0000 (0:00:03.837)       0:00:41.148 ******** \n=============================================================================== \nosism.commons.repository : Update package cache ------------------------ 17.43s\nInstall required packages (Debian) -------------------------------------- 7.29s\nGathers facts about hosts ----------------------------------------------- 3.84s\nCopy fact files --------------------------------------------------------- 3.80s\nosism.commons.repository : Force update of package cache ---------------- 1.55s\nCreate custom facts directory ------------------------------------------- 1.35s\nCopy fact file ---------------------------------------------------------- 1.35s\nosism.commons.repository : Copy ubuntu.sources file --------------------- 1.14s\nosism.commons.repository : Copy 99osism apt configuration --------------- 1.06s\nosism.commons.repository : Remove sources.list file --------------------- 0.50s\nCreate custom facts directory ------------------------------------------- 0.49s\nosism.commons.repository : Create /etc/apt/sources.list.d directory ----- 0.43s\nosism.commons.repository : Set repository_default fact to default value --- 0.20s\nosism.commons.repository : Set repositories to default ------------------ 0.19s\nosism.commons.repository : Include distribution specific repository tasks --- 0.12s\nosism.commons.repository : Include tasks for Ubuntu < 24.04 ------------- 0.11s\nosism.commons.repository : Gather variables for each operating system --- 0.10s\nInstall required packages (RedHat) -------------------------------------- 0.09s\n+ osism apply bootstrap\n\u001b[32m2026-03-21 00:25:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of bootstrap.\u001b[0m\n\u001b[32m2026-03-21 00:25:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask adc203e2-da6f-4ccb-9c93-88f3a7415320 (bootstrap) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 00:25:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task adc203e2-da6f-4ccb-9c93-88f3a7415320 (bootstrap) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on state bootstrap] ************************************\n\nTASK [Group hosts based on state bootstrap] ************************************\nSaturday 21 March 2026  00:25:38 +0000 (0:00:00.200)       0:00:00.200 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nSaturday 21 March 2026  00:25:38 +0000 (0:00:00.303)       0:00:00.503 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nPLAY [Gather facts for all hosts (if using --limit)] ***************************\n\nTASK [Gathers facts about hosts] ***********************************************\nSaturday 21 March 2026  00:25:43 +0000 (0:00:04.886)       0:00:05.390 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY [Apply bootstrap roles part 1] ********************************************\n\nTASK [osism.commons.hostname : Set hostname] ***********************************\nSaturday 21 March 2026  00:25:44 +0000 (0:00:00.449)       0:00:05.839 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.hostname : Copy /etc/hostname] *****************************\nSaturday 21 March 2026  00:25:45 +0000 (0:00:01.460)       0:00:07.300 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.hosts : Include type specific tasks] ***********************\nSaturday 21 March 2026  00:25:46 +0000 (0:00:01.249)       0:00:08.549 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/hosts/tasks/type-template.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.hosts : Copy /etc/hosts file] ******************************\nSaturday 21 March 2026  00:25:47 +0000 (0:00:00.267)       0:00:08.817 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.proxy : Include distribution specific tasks] ***************\nSaturday 21 March 2026  00:25:48 +0000 (0:00:01.489)       0:00:10.306 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/proxy/tasks/Debian-family.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.proxy : Configure proxy parameters for apt] ****************\nSaturday 21 March 2026  00:25:49 +0000 (0:00:00.297)       0:00:10.603 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.proxy : Set system wide settings in environment file] ******\nSaturday 21 March 2026  00:25:50 +0000 (0:00:01.030)       0:00:11.634 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.proxy : Remove system wide settings in environment file] ***\nSaturday 21 March 2026  00:25:50 +0000 (0:00:00.593)       0:00:12.227 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Check minimum and maximum number of name servers] ***\nSaturday 21 March 2026  00:25:51 +0000 (0:00:00.407)       0:00:12.635 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.resolvconf : Include resolvconf tasks] *********************\nSaturday 21 March 2026  00:25:51 +0000 (0:00:00.186)       0:00:12.822 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-resolv.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.resolvconf : Include distribution specific installation tasks] ***\nSaturday 21 March 2026  00:25:51 +0000 (0:00:00.284)       0:00:13.106 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf] ***\nSaturday 21 March 2026  00:25:51 +0000 (0:00:00.293)       0:00:13.400 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.resolvconf : Install package systemd-resolved] *************\nSaturday 21 March 2026  00:25:53 +0000 (0:00:01.308)       0:00:14.708 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf] *****\nSaturday 21 March 2026  00:25:53 +0000 (0:00:00.205)       0:00:14.914 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.resolvconf : Archive existing file /etc/resolv.conf] *******\nSaturday 21 March 2026  00:25:53 +0000 (0:00:00.589)       0:00:15.503 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf] ***\nSaturday 21 March 2026  00:25:54 +0000 (0:00:00.227)       0:00:15.730 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.resolvconf : Copy configuration files] *********************\nSaturday 21 March 2026  00:25:54 +0000 (0:00:00.679)       0:00:16.410 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.resolvconf : Start/enable systemd-resolved service] ********\nSaturday 21 March 2026  00:25:56 +0000 (0:00:01.185)       0:00:17.595 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.resolvconf : Include distribution specific configuration tasks] ***\nSaturday 21 March 2026  00:25:57 +0000 (0:00:01.017)       0:00:18.613 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.resolvconf : Restart systemd-resolved service] *************\nSaturday 21 March 2026  00:25:57 +0000 (0:00:00.330)       0:00:18.944 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.repository : Gather variables for each operating system] ***\nSaturday 21 March 2026  00:25:58 +0000 (0:00:01.584)       0:00:20.529 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Set repository_default fact to default value] ***\nSaturday 21 March 2026  00:25:59 +0000 (0:00:00.242)       0:00:20.771 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Set repositories to default] ******************\nSaturday 21 March 2026  00:25:59 +0000 (0:00:00.252)       0:00:21.024 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Include distribution specific repository tasks] ***\nSaturday 21 March 2026  00:25:59 +0000 (0:00:00.227)       0:00:21.252 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****\nSaturday 21 March 2026  00:25:59 +0000 (0:00:00.320)       0:00:21.572 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************\nSaturday 21 March 2026  00:26:00 +0000 (0:00:00.620)       0:00:22.193 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Copy 99osism apt configuration] ***************\nSaturday 21 March 2026  00:26:00 +0000 (0:00:00.216)       0:00:22.410 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Remove sources.list file] *********************\nSaturday 21 March 2026  00:26:01 +0000 (0:00:01.095)       0:00:23.506 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Copy ubuntu.sources file] *********************\nSaturday 21 March 2026  00:26:02 +0000 (0:00:00.620)       0:00:24.126 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.repository : Update package cache] *************************\nSaturday 21 March 2026  00:26:03 +0000 (0:00:01.204)       0:00:25.331 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.rsyslog : Gather variables for each operating system] *****\nSaturday 21 March 2026  00:26:20 +0000 (0:00:16.543)       0:00:41.874 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.rsyslog : Set rsyslog_user variable to default value] *****\nSaturday 21 March 2026  00:26:20 +0000 (0:00:00.250)       0:00:42.124 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.rsyslog : Set rsyslog_workdir variable to default value] ***\nSaturday 21 March 2026  00:26:20 +0000 (0:00:00.249)       0:00:42.374 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.rsyslog : Include distribution specific install tasks] ****\nSaturday 21 March 2026  00:26:21 +0000 (0:00:00.246)       0:00:42.620 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.rsyslog : Install rsyslog package] ************************\nSaturday 21 March 2026  00:26:21 +0000 (0:00:00.280)       0:00:42.901 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.rsyslog : Copy rsyslog.conf configuration file] ***********\nSaturday 21 March 2026  00:26:23 +0000 (0:00:02.151)       0:00:45.053 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.rsyslog : Manage rsyslog service] *************************\nSaturday 21 March 2026  00:26:24 +0000 (0:00:01.186)       0:00:46.239 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.rsyslog : Include fluentd tasks] **************************\nSaturday 21 March 2026  00:26:25 +0000 (0:00:00.978)       0:00:47.217 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/fluentd.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.rsyslog : Forward syslog message to local fluentd daemon] ***\nSaturday 21 March 2026  00:26:25 +0000 (0:00:00.324)       0:00:47.542 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.services.rsyslog : Include additional log server tasks] ************\nSaturday 21 March 2026  00:26:27 +0000 (0:00:01.196)       0:00:48.738 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.rsyslog : Include logrotate tasks] ************************\nSaturday 21 March 2026  00:26:27 +0000 (0:00:00.265)       0:00:49.004 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/logrotate.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.rsyslog : Ensure logrotate package is installed] **********\nSaturday 21 March 2026  00:26:27 +0000 (0:00:00.350)       0:00:49.354 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.rsyslog : Configure logrotate for rsyslog] ****************\nSaturday 21 March 2026  00:26:29 +0000 (0:00:01.931)       0:00:51.286 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.systohc : Install util-linux-extra package] ****************\nSaturday 21 March 2026  00:26:30 +0000 (0:00:01.223)       0:00:52.509 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.systohc : Sync hardware clock] *****************************\nSaturday 21 March 2026  00:26:42 +0000 (0:00:11.798)       0:01:04.308 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.configfs : Start sys-kernel-config mount] ******************\nSaturday 21 March 2026  00:26:43 +0000 (0:00:00.939)       0:01:05.247 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.packages : Gather variables for each operating system] *****\nSaturday 21 March 2026  00:26:44 +0000 (0:00:01.030)       0:01:06.278 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.packages : Set required_packages_distribution variable to default value] ***\nSaturday 21 March 2026  00:26:44 +0000 (0:00:00.218)       0:01:06.497 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.packages : Include distribution specific package tasks] ****\nSaturday 21 March 2026  00:26:45 +0000 (0:00:00.205)       0:01:06.702 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/packages/tasks/package-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.packages : Install needrestart package] ********************\nSaturday 21 March 2026  00:26:45 +0000 (0:00:00.259)       0:01:06.962 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.packages : Set needrestart mode] ***************************\nSaturday 21 March 2026  00:26:47 +0000 (0:00:01.953)       0:01:08.915 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.packages : Set apt_cache_valid_time variable to default value] ***\nSaturday 21 March 2026  00:26:48 +0000 (0:00:00.730)       0:01:09.646 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.packages : Update package cache] ***************************\nSaturday 21 March 2026  00:26:48 +0000 (0:00:00.304)       0:01:09.950 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.packages : Download upgrade packages] **********************\nSaturday 21 March 2026  00:26:49 +0000 (0:00:01.379)       0:01:11.329 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.packages : Upgrade packages] *******************************\nSaturday 21 March 2026  00:26:51 +0000 (0:00:02.211)       0:01:13.541 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.packages : Download required packages] *********************\nSaturday 21 March 2026  00:26:54 +0000 (0:00:02.865)       0:01:16.406 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.packages : Install required packages] **********************\nSaturday 21 March 2026  00:27:32 +0000 (0:00:37.683)       0:01:54.090 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.packages : Remove useless packages from the cache] *********\nSaturday 21 March 2026  00:28:43 +0000 (0:01:10.988)       0:03:05.079 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.packages : Remove dependencies that are no longer required] ***\nSaturday 21 March 2026  00:28:45 +0000 (0:00:01.862)       0:03:06.942 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.sysctl : Include sysctl tasks] *****************************\nSaturday 21 March 2026  00:28:57 +0000 (0:00:11.961)       0:03:18.903 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 => (item={'key': 'elasticsearch', 'value': [{'name': 'vm.max_map_count', 'value': 262144}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 => (item={'key': 'rabbitmq', 'value': [{'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}, {'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}, {'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}, {'name': 'net.core.wmem_max', 'value': 16777216}, {'name': 'net.core.rmem_max', 'value': 16777216}, {'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}, {'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}, {'name': 'net.core.somaxconn', 'value': 4096}, {'name': 'net.ipv4.tcp_syncookies', 'value': 0}, {'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 => (item={'key': 'generic', 'value': [{'name': 'vm.swappiness', 'value': 1}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 => (item={'key': 'compute', 'value': [{'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 => (item={'key': 'network', 'value': [{'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 => (item={'key': 'k3s_node', 'value': [{'name': 'fs.inotify.max_user_instances', 'value': 1024}]})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on elasticsearch] ***********\nSaturday 21 March 2026  00:28:57 +0000 (0:00:00.385)       0:03:19.289 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on rabbitmq] ****************\nSaturday 21 March 2026  00:28:58 +0000 (0:00:00.604)       0:03:19.893 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on generic] *****************\nSaturday 21 March 2026  00:29:04 +0000 (0:00:06.033)       0:03:25.927 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on compute] *****************\nSaturday 21 March 2026  00:29:05 +0000 (0:00:01.634)       0:03:27.561 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on network] *****************\nSaturday 21 March 2026  00:29:06 +0000 (0:00:00.599)       0:03:28.161 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on k3s_node] ****************\nSaturday 21 March 2026  00:29:07 +0000 (0:00:00.536)       0:03:28.698 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m\n\nTASK [osism.commons.limits : Include limits tasks] *****************************\nSaturday 21 March 2026  00:29:07 +0000 (0:00:00.717)       0:03:29.415 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.services : Populate service facts] *************************\nSaturday 21 March 2026  00:29:08 +0000 (0:00:00.299)       0:03:29.715 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.services : Check services] *********************************\nSaturday 21 March 2026  00:29:13 +0000 (0:00:05.164)       0:03:34.879 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.services : Start/enable required services] *****************\nSaturday 21 March 2026  00:29:13 +0000 (0:00:00.289)       0:03:35.168 ******** \n\u001b[0;32mok: [testbed-manager] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=cron)\u001b[0m\n\nTASK [osism.commons.motd : Include distribution specific configure tasks] ******\nSaturday 21 March 2026  00:29:14 +0000 (0:00:01.138)       0:03:36.307 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/motd/tasks/configure-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.motd : Remove update-motd package] *************************\nSaturday 21 March 2026  00:29:15 +0000 (0:00:00.407)       0:03:36.714 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.motd : Check if /etc/default/motd-news exists] *************\nSaturday 21 March 2026  00:29:16 +0000 (0:00:01.380)       0:03:38.094 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.motd : Disable the dynamic motd-news service] **************\nSaturday 21 March 2026  00:29:17 +0000 (0:00:00.661)       0:03:38.756 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.motd : Get all configuration files in /etc/pam.d] **********\nSaturday 21 March 2026  00:29:17 +0000 (0:00:00.676)       0:03:39.432 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.motd : Remove pam_motd.so rule] ****************************\nSaturday 21 March 2026  00:29:18 +0000 (0:00:00.636)       0:03:40.068 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774051521.6864767, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774051567.6236367, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774051572.2527146, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774051533.1850278, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774051532.6675546, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774051566.528907, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774051573.9143372, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\nTASK [osism.commons.motd : Copy motd file] *************************************\nSaturday 21 March 2026  00:29:19 +0000 (0:00:01.018)       0:03:41.087 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.motd : Copy issue file] ************************************\nSaturday 21 March 2026  00:29:20 +0000 (0:00:01.094)       0:03:42.181 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.motd : Copy issue.net file] ********************************\nSaturday 21 March 2026  00:29:21 +0000 (0:00:01.206)       0:03:43.387 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.motd : Configure SSH to print the motd] ********************\nSaturday 21 March 2026  00:29:23 +0000 (0:00:01.252)       0:03:44.640 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.motd : Configure SSH to not print the motd] ****************\nSaturday 21 March 2026  00:29:23 +0000 (0:00:00.258)       0:03:44.898 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.rng : Include distribution specific install tasks] ********\nSaturday 21 March 2026  00:29:24 +0000 (0:00:00.861)       0:03:45.760 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rng/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.rng : Install rng package] ********************************\nSaturday 21 March 2026  00:29:24 +0000 (0:00:00.411)       0:03:46.171 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.rng : Remove haveged package] *****************************\nSaturday 21 March 2026  00:29:33 +0000 (0:00:08.446)       0:03:54.618 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.rng : Manage rng service] *********************************\nSaturday 21 March 2026  00:29:34 +0000 (0:00:01.525)       0:03:56.144 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.cleanup : Gather variables for each operating system] ******\nSaturday 21 March 2026  00:29:35 +0000 (0:00:01.228)       0:03:57.372 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.cleanup : Set cleanup_packages_distribution variable to default value] ***\nSaturday 21 March 2026  00:29:36 +0000 (0:00:00.294)       0:03:57.667 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.cleanup : Set cleanup_services_distribution variable to default value] ***\nSaturday 21 March 2026  00:29:36 +0000 (0:00:00.290)       0:03:57.957 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.cleanup : Populate service facts] **************************\nSaturday 21 March 2026  00:29:36 +0000 (0:00:00.301)       0:03:58.259 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.cleanup : Include distribution specific timer tasks] *******\nSaturday 21 March 2026  00:29:41 +0000 (0:00:04.876)       0:04:03.135 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/timers-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.cleanup : Disable apt-daily timers] ************************\nSaturday 21 March 2026  00:29:41 +0000 (0:00:00.398)       0:04:03.533 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.cleanup : Include service tasks] ***************************\nSaturday 21 March 2026  00:29:42 +0000 (0:00:00.392)       0:04:03.926 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/services-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.cleanup : Cleanup services] ********************************\nSaturday 21 March 2026  00:29:42 +0000 (0:00:00.570)       0:04:04.496 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.cleanup : Include packages tasks] **************************\nSaturday 21 March 2026  00:29:43 +0000 (0:00:00.350)       0:04:04.847 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/packages-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.cleanup : Cleanup installed packages] **********************\nSaturday 21 March 2026  00:29:43 +0000 (0:00:00.456)       0:04:05.303 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.cleanup : Remove cloudinit package] ************************\nSaturday 21 March 2026  00:30:15 +0000 (0:00:31.716)       0:04:37.020 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.cleanup : Uninstall unattended-upgrades package] ***********\nSaturday 21 March 2026  00:30:24 +0000 (0:00:08.981)       0:04:46.002 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.cleanup : Remove useless packages from the cache] **********\nSaturday 21 March 2026  00:30:33 +0000 (0:00:08.684)       0:04:54.686 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.cleanup : Remove dependencies that are no longer required] ***\nSaturday 21 March 2026  00:30:34 +0000 (0:00:01.817)       0:04:56.504 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.cleanup : Include cloudinit tasks] *************************\nSaturday 21 March 2026  00:30:41 +0000 (0:00:06.582)       0:05:03.086 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/cloudinit.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.cleanup : Remove cloud-init configuration directory] *******\nSaturday 21 March 2026  00:30:41 +0000 (0:00:00.385)       0:05:03.472 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.timezone : Install tzdata package] *************************\nSaturday 21 March 2026  00:30:42 +0000 (0:00:00.740)       0:05:04.212 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.timezone : Set timezone to UTC] ****************************\nSaturday 21 March 2026  00:30:44 +0000 (0:00:01.786)       0:05:05.999 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.timezone : Create /etc/adjtime file] ***********************\nSaturday 21 March 2026  00:30:45 +0000 (0:00:00.737)       0:05:06.737 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.timezone : Ensure UTC in /etc/adjtime] *********************\nSaturday 21 March 2026  00:30:45 +0000 (0:00:00.271)       0:05:07.008 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Gather variables for each operating system] ******\nSaturday 21 March 2026  00:30:45 +0000 (0:00:00.382)       0:05:07.390 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Set docker_version variable to default value] ****\nSaturday 21 March 2026  00:30:46 +0000 (0:00:00.427)       0:05:07.818 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Set docker_cli_version variable to default value] ***\nSaturday 21 March 2026  00:30:46 +0000 (0:00:00.278)       0:05:08.097 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Print used docker version] ***********************\nSaturday 21 March 2026  00:30:46 +0000 (0:00:00.295)       0:05:08.393 ******** \n\u001b[0;32mok: [testbed-manager] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\nTASK [osism.services.docker : Print used docker cli version] *******************\nSaturday 21 March 2026  00:30:47 +0000 (0:00:00.256)       0:05:08.649 ******** \n\u001b[0;32mok: [testbed-manager] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\nTASK [osism.services.docker : Include block storage tasks] *********************\nSaturday 21 March 2026  00:30:47 +0000 (0:00:00.266)       0:05:08.916 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Include zram storage tasks] **********************\nSaturday 21 March 2026  00:30:47 +0000 (0:00:00.238)       0:05:09.154 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Include docker install tasks] ********************\nSaturday 21 March 2026  00:30:47 +0000 (0:00:00.230)       0:05:09.385 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/install-docker-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.docker : Remove old architecture-dependent repository] ****\nSaturday 21 March 2026  00:30:48 +0000 (0:00:00.387)       0:05:09.772 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Gather package facts] ****************************\nSaturday 21 March 2026  00:30:49 +0000 (0:00:00.932)       0:05:10.705 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [osism.services.docker : Check whether packages are installed that should not be installed] ***\nSaturday 21 March 2026  00:30:52 +0000 (0:00:03.169)       0:05:13.874 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Install apt-transport-https package] *************\nSaturday 21 March 2026  00:30:52 +0000 (0:00:00.581)       0:05:14.456 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [osism.services.docker : Add repository gpg key] **************************\nSaturday 21 March 2026  00:30:59 +0000 (0:00:07.032)       0:05:21.488 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Add repository] **********************************\nSaturday 21 March 2026  00:31:01 +0000 (0:00:01.158)       0:05:22.646 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Update package cache] ****************************\nSaturday 21 March 2026  00:31:09 +0000 (0:00:08.294)       0:05:30.940 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Pin docker package version] **********************\nSaturday 21 March 2026  00:31:12 +0000 (0:00:03.609)       0:05:34.550 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Pin docker-cli package version] ******************\nSaturday 21 March 2026  00:31:14 +0000 (0:00:01.339)       0:05:35.889 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Unlock containerd package] ***********************\nSaturday 21 March 2026  00:31:15 +0000 (0:00:01.380)       0:05:37.269 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Install containerd package] **********************\nSaturday 21 March 2026  00:31:16 +0000 (0:00:00.579)       0:05:37.849 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [osism.services.docker : Lock containerd package] *************************\nSaturday 21 March 2026  00:31:26 +0000 (0:00:10.532)       0:05:48.383 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Install docker-cli package] **********************\nSaturday 21 March 2026  00:31:27 +0000 (0:00:01.157)       0:05:49.540 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Install docker package] **************************\nSaturday 21 March 2026  00:31:37 +0000 (0:00:09.764)       0:05:59.305 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.services.docker : Unblock installation of python docker packages] ***\nSaturday 21 March 2026  00:31:49 +0000 (0:00:11.395)       0:06:10.701 ******** \n\u001b[0;32mok: [testbed-manager] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=python-docker)\u001b[0m\n\nTASK [osism.services.docker : Install python3 docker package] ******************\nSaturday 21 March 2026  00:31:50 +0000 (0:00:01.232)       0:06:11.933 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Install python3 docker package from Debian Sid] ***\nSaturday 21 March 2026  00:31:51 +0000 (0:00:00.660)       0:06:12.594 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Remove python docker packages (install python bindings from pip)] ***\nSaturday 21 March 2026  00:31:55 +0000 (0:00:04.047)       0:06:16.641 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Block installation of python docker packages (install python bindings from pip)] ***\nSaturday 21 March 2026  00:31:55 +0000 (0:00:00.505)       0:06:17.146 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Install python3-pip package (install python bindings from pip)] ***\nSaturday 21 March 2026  00:31:56 +0000 (0:00:00.542)       0:06:17.689 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Install docker packages (install python bindings from pip)] ***\nSaturday 21 March 2026  00:31:56 +0000 (0:00:00.479)       0:06:18.168 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Install packages required by docker login] *******\nSaturday 21 March 2026  00:31:57 +0000 (0:00:00.624)       0:06:18.793 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Ensure that some packages are not installed] *****\nSaturday 21 March 2026  00:31:57 +0000 (0:00:00.525)       0:06:19.319 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Include config tasks] ****************************\nSaturday 21 March 2026  00:31:59 +0000 (0:00:01.941)       0:06:21.260 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/config.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.docker : Create plugins directory] ************************\nSaturday 21 March 2026  00:32:00 +0000 (0:00:00.827)       0:06:22.088 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Create systemd overlay directory] ****************\nSaturday 21 March 2026  00:32:01 +0000 (0:00:01.086)       0:06:23.174 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Copy systemd overlay file] ***********************\nSaturday 21 March 2026  00:32:02 +0000 (0:00:00.853)       0:06:24.028 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Reload systemd daemon if systemd overlay file is changed] ***\nSaturday 21 March 2026  00:32:03 +0000 (0:00:01.345)       0:06:25.373 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Copy limits configuration file] ******************\nSaturday 21 March 2026  00:32:05 +0000 (0:00:01.372)       0:06:26.745 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Copy daemon.json configuration file] *************\nSaturday 21 March 2026  00:32:06 +0000 (0:00:01.352)       0:06:28.097 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Include service tasks] ***************************\nSaturday 21 March 2026  00:32:08 +0000 (0:00:01.643)       0:06:29.740 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/service.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.docker : Reload systemd daemon] ***************************\nSaturday 21 March 2026  00:32:08 +0000 (0:00:00.850)       0:06:30.591 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Manage service] **********************************\nSaturday 21 March 2026  00:32:10 +0000 (0:00:01.381)       0:06:31.973 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Manage docker socket service] ********************\nSaturday 21 March 2026  00:32:11 +0000 (0:00:01.283)       0:06:33.256 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Manage containerd service] ***********************\nSaturday 21 March 2026  00:32:12 +0000 (0:00:01.149)       0:06:34.406 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Include bootstrap tasks] *************************\nSaturday 21 March 2026  00:32:13 +0000 (0:00:01.100)       0:06:35.506 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/bootstrap.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.docker : Flush handlers] **********************************\nSaturday 21 March 2026  00:32:14 +0000 (0:00:00.918)       0:06:36.425 ******** \n\nTASK [osism.services.docker : Flush handlers] **********************************\nSaturday 21 March 2026  00:32:14 +0000 (0:00:00.046)       0:06:36.472 ******** \n\nTASK [osism.services.docker : Flush handlers] **********************************\nSaturday 21 March 2026  00:32:15 +0000 (0:00:00.242)       0:06:36.714 ******** \n\nTASK [osism.services.docker : Flush handlers] **********************************\nSaturday 21 March 2026  00:32:15 +0000 (0:00:00.041)       0:06:36.756 ******** \n\nTASK [osism.services.docker : Flush handlers] **********************************\nSaturday 21 March 2026  00:32:15 +0000 (0:00:00.041)       0:06:36.797 ******** \n\nTASK [osism.services.docker : Flush handlers] **********************************\nSaturday 21 March 2026  00:32:15 +0000 (0:00:00.051)       0:06:36.848 ******** \n\nTASK [osism.services.docker : Flush handlers] **********************************\nSaturday 21 March 2026  00:32:15 +0000 (0:00:00.041)       0:06:36.890 ******** \n\nRUNNING HANDLER [osism.commons.repository : Force update of package cache] *****\nSaturday 21 March 2026  00:32:15 +0000 (0:00:00.042)       0:06:36.932 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [osism.services.rsyslog : Restart rsyslog service] *************\nSaturday 21 March 2026  00:32:16 +0000 (0:00:01.336)       0:06:38.269 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [osism.services.rsyslog : Restart logrotate service] ***********\nSaturday 21 March 2026  00:32:17 +0000 (0:00:01.312)       0:06:39.581 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [osism.services.docker : Restart docker service] ***************\nSaturday 21 March 2026  00:32:19 +0000 (0:00:01.191)       0:06:40.773 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nRUNNING HANDLER [osism.services.docker : Wait after docker service restart] ****\nSaturday 21 March 2026  00:32:21 +0000 (0:00:02.508)       0:06:43.281 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [osism.services.docker : Add user to docker group] ************************\nSaturday 21 March 2026  00:32:21 +0000 (0:00:00.103)       0:06:43.384 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Log into private registry and force re-authorization] ***\nSaturday 21 March 2026  00:32:22 +0000 (0:00:01.180)       0:06:44.565 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Include facts tasks] *****************************\nSaturday 21 March 2026  00:32:23 +0000 (0:00:00.502)       0:06:45.067 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/facts.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.docker : Create facts directory] **************************\nSaturday 21 March 2026  00:32:24 +0000 (0:00:00.890)       0:06:45.958 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Copy docker fact files] **************************\nSaturday 21 March 2026  00:32:25 +0000 (0:00:01.205)       0:06:47.163 ******** \n\u001b[0;32mok: [testbed-manager] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=docker_containers)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=docker_images)\u001b[0m\n\nTASK [osism.commons.docker_compose : This install type is not supported] *******\nSaturday 21 March 2026  00:32:28 +0000 (0:00:02.614)       0:06:49.778 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.docker_compose : Include distribution specific install tasks] ***\nSaturday 21 March 2026  00:32:28 +0000 (0:00:00.485)       0:06:50.264 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/docker_compose/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.docker_compose : Remove docker-compose apt preferences file] ***\nSaturday 21 March 2026  00:32:29 +0000 (0:00:00.926)       0:06:51.191 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.docker_compose : Get checksum of docker-compose file] ******\nSaturday 21 March 2026  00:32:30 +0000 (0:00:00.868)       0:06:52.059 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.docker_compose : Remove docker-compose binary] *************\nSaturday 21 March 2026  00:32:31 +0000 (0:00:00.832)       0:06:52.892 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.docker_compose : Uninstall docker-compose package] *********\nSaturday 21 March 2026  00:32:31 +0000 (0:00:00.526)       0:06:53.419 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.docker_compose : Copy docker-compose script] ***************\nSaturday 21 March 2026  00:32:33 +0000 (0:00:01.738)       0:06:55.157 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.docker_compose : Install docker-compose-plugin package] ****\nSaturday 21 March 2026  00:32:34 +0000 (0:00:00.779)       0:06:55.937 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.docker_compose : Copy osism.target systemd file] ***********\nSaturday 21 March 2026  00:32:42 +0000 (0:00:07.891)       0:07:03.828 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.docker_compose : Enable osism.target] **********************\nSaturday 21 March 2026  00:32:43 +0000 (0:00:01.348)       0:07:05.177 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.docker_compose : Copy docker-compose systemd unit file] ****\nSaturday 21 March 2026  00:32:45 +0000 (0:00:01.760)       0:07:06.937 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.facts : Create custom facts directory] *********************\nSaturday 21 March 2026  00:32:47 +0000 (0:00:01.887)       0:07:08.825 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.facts : Copy fact files] ***********************************\nSaturday 21 March 2026  00:32:48 +0000 (0:00:00.848)       0:07:09.673 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.chrony : Check minimum and maximum number of servers] *****\nSaturday 21 March 2026  00:32:48 +0000 (0:00:00.763)       0:07:10.436 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.chrony : Gather variables for each operating system] ******\nSaturday 21 March 2026  00:32:49 +0000 (0:00:00.674)       0:07:11.111 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.chrony : Set chrony_conf_file variable to default value] ***\nSaturday 21 March 2026  00:32:50 +0000 (0:00:00.513)       0:07:11.624 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.chrony : Set chrony_key_file variable to default value] ***\nSaturday 21 March 2026  00:32:50 +0000 (0:00:00.518)       0:07:12.143 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.chrony : Populate service facts] **************************\nSaturday 21 March 2026  00:32:51 +0000 (0:00:00.500)       0:07:12.643 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.chrony : Manage timesyncd service] ************************\nSaturday 21 March 2026  00:32:56 +0000 (0:00:05.565)       0:07:18.208 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.chrony : Include distribution specific install tasks] *****\nSaturday 21 March 2026  00:32:57 +0000 (0:00:00.673)       0:07:18.881 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.chrony : Install package] *********************************\nSaturday 21 March 2026  00:32:58 +0000 (0:00:00.794)       0:07:19.676 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.chrony : Manage chrony service] ***************************\nSaturday 21 March 2026  00:33:00 +0000 (0:00:01.952)       0:07:21.629 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.chrony : Check if configuration file exists] **************\nSaturday 21 March 2026  00:33:01 +0000 (0:00:01.444)       0:07:23.073 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.chrony : Copy configuration file] *************************\nSaturday 21 March 2026  00:33:02 +0000 (0:00:00.910)       0:07:23.983 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\nTASK [osism.services.lldpd : Include distribution specific install tasks] ******\nSaturday 21 March 2026  00:33:04 +0000 (0:00:01.678)       0:07:25.662 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/lldpd/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.lldpd : Install lldpd package] ****************************\nSaturday 21 March 2026  00:33:05 +0000 (0:00:00.938)       0:07:26.601 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.services.lldpd : Manage lldpd service] *****************************\nSaturday 21 March 2026  00:33:14 +0000 (0:00:09.686)       0:07:36.287 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [osism.commons.docker_compose : Reload systemd daemon] *********\nSaturday 21 March 2026  00:33:16 +0000 (0:00:01.741)       0:07:38.029 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nRUNNING HANDLER [osism.services.chrony : Restart chrony service] ***************\nSaturday 21 March 2026  00:33:17 +0000 (0:00:01.557)       0:07:39.587 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY [Apply bootstrap role part 2] *********************************************\n\nTASK [Include hardening role] **************************************************\nSaturday 21 March 2026  00:33:19 +0000 (0:00:01.254)       0:07:40.841 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY [Apply bootstrap roles part 3] ********************************************\n\nTASK [osism.services.journald : Copy configuration file] ***********************\nSaturday 21 March 2026  00:33:19 +0000 (0:00:00.512)       0:07:41.354 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.journald : Manage journald service] ***********************\nSaturday 21 March 2026  00:33:21 +0000 (0:00:01.387)       0:07:42.741 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Include auditd role] *****************************************************\nSaturday 21 March 2026  00:33:22 +0000 (0:00:01.627)       0:07:44.369 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Include smartd role] *****************************************************\nSaturday 21 March 2026  00:33:23 +0000 (0:00:00.459)       0:07:44.829 ******** \n\u001b[0;36mincluded: osism.services.smartd for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.smartd : Include distribution specific install tasks] *****\nSaturday 21 March 2026  00:33:24 +0000 (0:00:00.791)       0:07:45.621 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/smartd/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.smartd : Install smartmontools package] *******************\nSaturday 21 March 2026  00:33:24 +0000 (0:00:00.921)       0:07:46.542 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.services.smartd : Create /var/log/smartd directory] ****************\nSaturday 21 March 2026  00:33:33 +0000 (0:00:09.044)       0:07:55.587 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.smartd : Copy smartmontools configuration file] ***********\nSaturday 21 March 2026  00:33:34 +0000 (0:00:00.838)       0:07:56.426 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.smartd : Manage smartd service] ***************************\nSaturday 21 March 2026  00:33:36 +0000 (0:00:01.372)       0:07:57.798 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [osism.services.journald : Restart journald service] ***********\nSaturday 21 March 2026  00:33:38 +0000 (0:00:01.881)       0:07:59.679 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [osism.services.smartd : Restart smartd service] ***************\nSaturday 21 March 2026  00:33:39 +0000 (0:00:01.202)       0:08:00.882 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY [Set state bootstrap] *****************************************************\n\nTASK [Set osism.bootstrap.status fact] *****************************************\nSaturday 21 March 2026  00:33:40 +0000 (0:00:01.100)       0:08:01.983 ******** \n\u001b[0;36mincluded: osism.commons.state for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.state : Create custom facts directory] *********************\nSaturday 21 March 2026  00:33:41 +0000 (0:00:01.091)       0:08:03.075 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.state : Write state into file] *****************************\nSaturday 21 March 2026  00:33:42 +0000 (0:00:00.832)       0:08:03.907 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Set osism.bootstrap.timestamp fact] **************************************\nSaturday 21 March 2026  00:33:43 +0000 (0:00:01.245)       0:08:05.153 ******** \n\u001b[0;36mincluded: osism.commons.state for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.state : Create custom facts directory] *********************\nSaturday 21 March 2026  00:33:44 +0000 (0:00:00.822)       0:08:05.976 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.state : Write state into file] *****************************\nSaturday 21 March 2026  00:33:45 +0000 (0:00:00.831)       0:08:06.807 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=168 \u001b[0m \u001b[0;33mchanged=40  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=42  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=69  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=37  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=69  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=69  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=65  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=37  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=65  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=37  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=65  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=37  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:33:46 +0000 (0:00:01.312)       0:08:08.120 ******** \n=============================================================================== \nosism.commons.packages : Install required packages --------------------- 70.99s\nosism.commons.packages : Download required packages -------------------- 37.68s\nosism.commons.cleanup : Cleanup installed packages --------------------- 31.72s\nosism.commons.repository : Update package cache ------------------------ 16.54s\nosism.commons.packages : Remove dependencies that are no longer required -- 11.96s\nosism.commons.systohc : Install util-linux-extra package --------------- 11.80s\nosism.services.docker : Install docker package ------------------------- 11.40s\nosism.services.docker : Install containerd package --------------------- 10.53s\nosism.services.docker : Install docker-cli package ---------------------- 9.77s\nosism.services.lldpd : Install lldpd package ---------------------------- 9.69s\nosism.services.smartd : Install smartmontools package ------------------- 9.04s\nosism.commons.cleanup : Remove cloudinit package ------------------------ 8.98s\nosism.commons.cleanup : Uninstall unattended-upgrades package ----------- 8.68s\nosism.services.rng : Install rng package -------------------------------- 8.45s\nosism.services.docker : Add repository ---------------------------------- 8.29s\nosism.commons.docker_compose : Install docker-compose-plugin package ---- 7.89s\nosism.services.docker : Install apt-transport-https package ------------- 7.03s\nosism.commons.cleanup : Remove dependencies that are no longer required --- 6.58s\nosism.commons.sysctl : Set sysctl parameters on rabbitmq ---------------- 6.03s\nosism.services.chrony : Populate service facts -------------------------- 5.57s\n+ osism apply fail2ban\n\u001b[32m2026-03-21 00:33:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of fail2ban.\u001b[0m\n\u001b[32m2026-03-21 00:33:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 647b7b4d-9406-4744-8f65-bebc9111ce84 (fail2ban) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 00:33:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 647b7b4d-9406-4744-8f65-bebc9111ce84 (fail2ban) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role fail2ban] *****************************************************\n\nTASK [osism.services.fail2ban : Include distribution specific install tasks] ***\nSaturday 21 March 2026  00:34:02 +0000 (0:00:00.330)       0:00:00.330 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/fail2ban/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.fail2ban : Install fail2ban package] **********************\nSaturday 21 March 2026  00:34:03 +0000 (0:00:01.113)       0:00:01.443 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.services.fail2ban : Copy configuration files] **********************\nSaturday 21 March 2026  00:34:15 +0000 (0:00:11.756)       0:00:13.200 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.fail2ban : Manage fail2ban service] ***********************\nSaturday 21 March 2026  00:34:16 +0000 (0:00:01.613)       0:00:14.814 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.fail2ban : Reload fail2ban configuration] *****************\nSaturday 21 March 2026  00:34:18 +0000 (0:00:01.244)       0:00:16.058 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:34:19 +0000 (0:00:01.520)       0:00:17.578 ******** \n=============================================================================== \nosism.services.fail2ban : Install fail2ban package --------------------- 11.76s\nosism.services.fail2ban : Copy configuration files ---------------------- 1.61s\nosism.services.fail2ban : Reload fail2ban configuration ----------------- 1.52s\nosism.services.fail2ban : Manage fail2ban service ----------------------- 1.24s\nosism.services.fail2ban : Include distribution specific install tasks --- 1.11s\n+ [[ -e /etc/redhat-release ]]\n+ osism apply network\n\u001b[32m2026-03-21 00:34:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of network.\u001b[0m\n\u001b[32m2026-03-21 00:34:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7cf0a656-9b93-4292-bfc5-7a3d9f2a5b10 (network) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 00:34:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 7cf0a656-9b93-4292-bfc5-7a3d9f2a5b10 (network) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role network] ******************************************************\n\nTASK [osism.commons.network : Gather variables for each operating system] ******\nSaturday 21 March 2026  00:34:34 +0000 (0:00:00.295)       0:00:00.295 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Include type specific tasks] *********************\nSaturday 21 March 2026  00:34:34 +0000 (0:00:00.535)       0:00:00.831 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/netplan-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.network : Install required packages] ***********************\nSaturday 21 March 2026  00:34:35 +0000 (0:00:01.024)       0:00:01.856 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove ifupdown package] *************************\nSaturday 21 March 2026  00:34:38 +0000 (0:00:02.491)       0:00:04.347 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Create required directories] *********************\nSaturday 21 March 2026  00:34:40 +0000 (0:00:01.599)       0:00:05.946 ******** \n\u001b[0;32mok: [testbed-manager] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/etc/netplan)\u001b[0m\n\nTASK [osism.commons.network : Prepare netplan configuration template] **********\nSaturday 21 March 2026  00:34:41 +0000 (0:00:01.062)       0:00:07.009 ******** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\nTASK [osism.commons.network : Copy netplan configuration] **********************\nSaturday 21 March 2026  00:34:44 +0000 (0:00:02.968)       0:00:09.978 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove netplan configuration template] ***********\nSaturday 21 March 2026  00:34:45 +0000 (0:00:01.492)       0:00:11.470 ******** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [osism.commons.network : Check if path for interface file exists] *********\nSaturday 21 March 2026  00:34:47 +0000 (0:00:01.657)       0:00:13.128 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Copy interfaces file] ****************************\nSaturday 21 March 2026  00:34:48 +0000 (0:00:00.892)       0:00:14.020 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Install package networkd-dispatcher] *************\nSaturday 21 March 2026  00:34:48 +0000 (0:00:00.636)       0:00:14.657 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Copy dispatcher scripts] *************************\nSaturday 21 March 2026  00:34:50 +0000 (0:00:02.003)       0:00:16.660 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'src': '/opt/configuration/network/iptables.sh', 'dest': 'routable.d/iptables.sh'})\u001b[0m\n\nTASK [osism.commons.network : Manage service networkd-dispatcher] **************\nSaturday 21 March 2026  00:34:51 +0000 (0:00:00.829)       0:00:17.490 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Include cleanup tasks] ***************************\nSaturday 21 March 2026  00:34:53 +0000 (0:00:01.550)       0:00:19.040 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/cleanup-netplan.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.network : List existing configuration files] ***************\nSaturday 21 March 2026  00:34:54 +0000 (0:00:01.213)       0:00:20.253 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Set network_configured_files fact] ***************\nSaturday 21 March 2026  00:34:55 +0000 (0:00:01.122)       0:00:21.375 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove unused configuration files] ***************\nSaturday 21 March 2026  00:34:56 +0000 (0:00:00.697)       0:00:22.073 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\nTASK [osism.commons.network : Include dummy interfaces] ************************\nSaturday 21 March 2026  00:34:57 +0000 (0:00:00.941)       0:00:23.015 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Include vxlan interfaces] ************************\nSaturday 21 March 2026  00:34:57 +0000 (0:00:00.698)       0:00:23.713 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/vxlan-interfaces.yml for testbed-node-1, testbed-node-0, testbed-manager, testbed-node-4, testbed-node-5, testbed-node-3, testbed-node-2\u001b[0m\n\nTASK [osism.commons.network : Create systemd networkd netdev files] ************\nSaturday 21 March 2026  00:35:01 +0000 (0:00:03.941)       0:00:27.655 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.5', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'addresses': ['192.168.112.5/20']}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.13', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': []}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.11', 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': []}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.10', 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': []}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.5', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'addresses': ['192.168.128.5/20']}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.12', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': []}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.14', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'addresses': []}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.13', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': ['192.168.128.13/20']}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.15', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'addresses': []}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.11', 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': ['192.168.128.11/20']}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.10', 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': ['192.168.128.10/20']}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.12', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': ['192.168.128.12/20']}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.14', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'addresses': ['192.168.128.14/20']}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.15', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'addresses': ['192.168.128.15/20']}})\u001b[0m\n\nTASK [osism.commons.network : Create systemd networkd network files] ***********\nSaturday 21 March 2026  00:35:07 +0000 (0:00:05.363)       0:00:33.018 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.5', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'addresses': ['192.168.112.5/20']}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.10', 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': []}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.11', 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': []}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.14', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'addresses': []}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.5', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'addresses': ['192.168.128.5/20']}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.12', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': []}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.15', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'addresses': []}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.10', 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': ['192.168.128.10/20']}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.13', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': []}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.11', 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': ['192.168.128.11/20']}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.14', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'addresses': ['192.168.128.14/20']}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.12', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': ['192.168.128.12/20']}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.15', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'addresses': ['192.168.128.15/20']}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.13', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': ['192.168.128.13/20']}})\u001b[0m\n\nTASK [osism.commons.network : Include networkd cleanup tasks] ******************\nSaturday 21 March 2026  00:35:12 +0000 (0:00:05.248)       0:00:38.267 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/cleanup-networkd.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.network : List existing configuration files] ***************\nSaturday 21 March 2026  00:35:13 +0000 (0:00:01.037)       0:00:39.305 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove unused configuration files] ***************\nSaturday 21 March 2026  00:35:14 +0000 (0:00:00.989)       0:00:40.294 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Include network extra init] **********************\nSaturday 21 March 2026  00:35:15 +0000 (0:00:00.645)       0:00:40.940 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/network-extra-init.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.network : Deploy network-extra-init script] ****************\nSaturday 21 March 2026  00:35:16 +0000 (0:00:01.054)       0:00:41.994 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Deploy network-extra-init systemd service] *******\nSaturday 21 March 2026  00:35:16 +0000 (0:00:00.630)       0:00:42.625 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Enable and start network-extra-init service] *****\nSaturday 21 March 2026  00:35:17 +0000 (0:00:00.521)       0:00:43.146 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Disable and stop network-extra-init service] *****\nSaturday 21 March 2026  00:35:17 +0000 (0:00:00.642)       0:00:43.789 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove network-extra-init systemd service] *******\nSaturday 21 March 2026  00:35:19 +0000 (0:00:01.469)       0:00:45.258 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove network-extra-init script] ****************\nSaturday 21 March 2026  00:35:20 +0000 (0:00:01.043)       0:00:46.301 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [osism.commons.network : Reload systemd-networkd] **************\nSaturday 21 March 2026  00:35:22 +0000 (0:00:02.139)       0:00:48.441 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [osism.commons.network : Netplan configuration changed] ********\nSaturday 21 March 2026  00:35:23 +0000 (0:00:00.602)       0:00:49.043 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=25  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:35:23 +0000 (0:00:00.671)       0:00:49.715 ******** \n=============================================================================== \nosism.commons.network : Create systemd networkd netdev files ------------ 5.36s\nosism.commons.network : Create systemd networkd network files ----------- 5.25s\nosism.commons.network : Include vxlan interfaces ------------------------ 3.94s\nosism.commons.network : Prepare netplan configuration template ---------- 2.97s\nosism.commons.network : Install required packages ----------------------- 2.49s\nosism.commons.network : Remove network-extra-init script ---------------- 2.14s\nosism.commons.network : Install package networkd-dispatcher ------------- 2.00s\nosism.commons.network : Remove netplan configuration template ----------- 1.66s\nosism.commons.network : Remove ifupdown package ------------------------- 1.60s\nosism.commons.network : Manage service networkd-dispatcher -------------- 1.55s\nosism.commons.network : Copy netplan configuration ---------------------- 1.49s\nosism.commons.network : Disable and stop network-extra-init service ----- 1.47s\nosism.commons.network : Include cleanup tasks --------------------------- 1.21s\nosism.commons.network : List existing configuration files --------------- 1.12s\nosism.commons.network : Create required directories --------------------- 1.06s\nosism.commons.network : Include network extra init ---------------------- 1.05s\nosism.commons.network : Remove network-extra-init systemd service ------- 1.04s\nosism.commons.network : Include networkd cleanup tasks ------------------ 1.04s\nosism.commons.network : Include type specific tasks --------------------- 1.02s\nosism.commons.network : List existing configuration files --------------- 0.99s\n+ osism apply wireguard\n\u001b[32m2026-03-21 00:35:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of wireguard.\u001b[0m\n\u001b[32m2026-03-21 00:35:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 101272ea-a17c-41c2-a57e-76286cdec110 (wireguard) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 00:35:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 101272ea-a17c-41c2-a57e-76286cdec110 (wireguard) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role wireguard] ****************************************************\n\nTASK [osism.services.wireguard : Install iptables package] *********************\nSaturday 21 March 2026  00:35:38 +0000 (0:00:00.330)       0:00:00.330 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Install wireguard package] ********************\nSaturday 21 March 2026  00:35:40 +0000 (0:00:01.760)       0:00:02.090 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Create public and private key - server] *******\nSaturday 21 March 2026  00:35:46 +0000 (0:00:06.105)       0:00:08.196 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Create preshared key] *************************\nSaturday 21 March 2026  00:35:47 +0000 (0:00:00.522)       0:00:08.718 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Get preshared key] ****************************\nSaturday 21 March 2026  00:35:47 +0000 (0:00:00.419)       0:00:09.138 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Get public key - server] **********************\nSaturday 21 March 2026  00:35:48 +0000 (0:00:00.559)       0:00:09.698 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Get private key - server] *********************\nSaturday 21 March 2026  00:35:48 +0000 (0:00:00.399)       0:00:10.097 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Copy wg0.conf configuration file] *************\nSaturday 21 March 2026  00:35:49 +0000 (0:00:00.406)       0:00:10.505 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Copy client configuration files] **************\nSaturday 21 March 2026  00:35:50 +0000 (0:00:01.157)       0:00:11.663 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Manage wg-quick@wg0.service service] **********\nSaturday 21 March 2026  00:35:51 +0000 (0:00:00.914)       0:00:12.577 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.wireguard : Restart wg0 service] ***************\nSaturday 21 March 2026  00:35:53 +0000 (0:00:01.886)       0:00:14.464 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=11  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:35:53 +0000 (0:00:00.868)       0:00:15.333 ******** \n=============================================================================== \nosism.services.wireguard : Install wireguard package -------------------- 6.11s\nosism.services.wireguard : Manage wg-quick@wg0.service service ---------- 1.89s\nosism.services.wireguard : Install iptables package --------------------- 1.76s\nosism.services.wireguard : Copy wg0.conf configuration file ------------- 1.16s\nosism.services.wireguard : Copy client configuration files -------------- 0.91s\nosism.services.wireguard : Restart wg0 service -------------------------- 0.87s\nosism.services.wireguard : Get preshared key ---------------------------- 0.56s\nosism.services.wireguard : Create public and private key - server ------- 0.52s\nosism.services.wireguard : Create preshared key ------------------------- 0.42s\nosism.services.wireguard : Get private key - server --------------------- 0.41s\nosism.services.wireguard : Get public key - server ---------------------- 0.40s\n+ sh -c /opt/configuration/scripts/prepare-wireguard-configuration.sh\n  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current\n                                 Dload  Upload   Total   Spent    Left  Speed\n\r  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0\r100    15  100    15    0     0    202      0 --:--:-- --:--:-- --:--:--   205\n+ osism apply --environment custom workarounds\n\u001b[32m2026-03-21 00:35:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play workarounds in environment custom\u001b[0m\n\u001b[32m2026-03-21 00:36:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of workarounds.\u001b[0m\n\u001b[32m2026-03-21 00:36:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 693fe854-71d3-49da-b5ea-134491721a6f (workarounds) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 00:36:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 693fe854-71d3-49da-b5ea-134491721a6f (workarounds) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on virtualization_role] ********************************\nSaturday 21 March 2026  00:36:08 +0000 (0:00:00.175)       0:00:00.175 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=virtualization_role_guest)\u001b[0m\n\nPLAY [Apply netplan configuration on the manager node] *************************\n\nTASK [Apply netplan configuration] *********************************************\nSaturday 21 March 2026  00:36:09 +0000 (0:00:00.778)       0:00:00.953 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY [Apply netplan configuration on all other nodes] **************************\n\nTASK [Apply netplan configuration] *********************************************\nSaturday 21 March 2026  00:36:12 +0000 (0:00:02.584)       0:00:03.537 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Add custom CA certificates to non-manager nodes] *************************\n\nTASK [Copy custom CA certificates] *********************************************\nSaturday 21 March 2026  00:36:14 +0000 (0:00:02.415)       0:00:05.953 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\nTASK [Run update-ca-certificates] **********************************************\nSaturday 21 March 2026  00:36:15 +0000 (0:00:01.390)       0:00:07.343 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Run update-ca-trust] *****************************************************\nSaturday 21 March 2026  00:36:19 +0000 (0:00:03.938)       0:00:11.282 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY [Add a workaround service] ************************************************\n\nTASK [Copy workarounds.sh scripts] *********************************************\nSaturday 21 March 2026  00:36:20 +0000 (0:00:00.546)       0:00:11.828 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Copy workarounds systemd unit file] **************************************\nSaturday 21 March 2026  00:36:22 +0000 (0:00:01.760)       0:00:13.589 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Reload systemd daemon] ***************************************************\nSaturday 21 March 2026  00:36:23 +0000 (0:00:01.489)       0:00:15.079 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Enable workarounds.service (Debian)] *************************************\nSaturday 21 March 2026  00:36:25 +0000 (0:00:01.604)       0:00:16.683 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Enable and start workarounds.service (RedHat)] ***************************\nSaturday 21 March 2026  00:36:26 +0000 (0:00:01.567)       0:00:18.250 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY [On Ubuntu 24.04 install python3-docker from Debian Sid] ******************\n\nTASK [Install python3-docker] **************************************************\nSaturday 21 March 2026  00:36:27 +0000 (0:00:00.739)       0:00:18.990 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:36:30 +0000 (0:00:02.629)       0:00:21.619 ******** \n=============================================================================== \nRun update-ca-certificates ---------------------------------------------- 3.94s\nInstall python3-docker -------------------------------------------------- 2.63s\nApply netplan configuration --------------------------------------------- 2.58s\nApply netplan configuration --------------------------------------------- 2.42s\nCopy workarounds.sh scripts --------------------------------------------- 1.76s\nReload systemd daemon --------------------------------------------------- 1.60s\nEnable workarounds.service (Debian) ------------------------------------- 1.57s\nCopy workarounds systemd unit file -------------------------------------- 1.49s\nCopy custom CA certificates --------------------------------------------- 1.39s\nGroup hosts based on virtualization_role -------------------------------- 0.78s\nEnable and start workarounds.service (RedHat) --------------------------- 0.74s\nRun update-ca-trust ----------------------------------------------------- 0.55s\n+ osism apply reboot -l testbed-nodes -e ireallymeanit=yes\n\u001b[32m2026-03-21 00:36:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of reboot.\u001b[0m\n\u001b[32m2026-03-21 00:36:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8dda0dcf-3289-4ca7-9fdd-cfa365aeff77 (reboot) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 00:36:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 8dda0dcf-3289-4ca7-9fdd-cfa365aeff77 (reboot) has been started and output is visible here.\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nSaturday 21 March 2026  00:36:44 +0000 (0:00:00.223)       0:00:00.223 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nSaturday 21 March 2026  00:36:44 +0000 (0:00:00.126)       0:00:00.350 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nSaturday 21 March 2026  00:36:46 +0000 (0:00:01.241)       0:00:01.591 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nSaturday 21 March 2026  00:36:46 +0000 (0:00:00.099)       0:00:01.690 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nSaturday 21 March 2026  00:36:46 +0000 (0:00:00.086)       0:00:01.777 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nSaturday 21 March 2026  00:36:47 +0000 (0:00:01.009)       0:00:02.787 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nSaturday 21 March 2026  00:36:47 +0000 (0:00:00.114)       0:00:02.901 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nSaturday 21 March 2026  00:36:47 +0000 (0:00:00.094)       0:00:02.996 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nSaturday 21 March 2026  00:36:48 +0000 (0:00:01.035)       0:00:04.031 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nSaturday 21 March 2026  00:36:48 +0000 (0:00:00.109)       0:00:04.141 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nSaturday 21 March 2026  00:36:48 +0000 (0:00:00.094)       0:00:04.235 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nSaturday 21 March 2026  00:36:49 +0000 (0:00:01.024)       0:00:05.259 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nSaturday 21 March 2026  00:36:49 +0000 (0:00:00.109)       0:00:05.369 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nSaturday 21 March 2026  00:36:50 +0000 (0:00:00.203)       0:00:05.573 ******** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nSaturday 21 March 2026  00:36:51 +0000 (0:00:01.000)       0:00:06.573 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nSaturday 21 March 2026  00:36:51 +0000 (0:00:00.111)       0:00:06.685 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nSaturday 21 March 2026  00:36:51 +0000 (0:00:00.096)       0:00:06.781 ******** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nSaturday 21 March 2026  00:36:52 +0000 (0:00:00.999)       0:00:07.781 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:36:52 +0000 (0:00:00.033)       0:00:07.814 ******** \n=============================================================================== \nReboot system - do not wait for the reboot to complete ------------------ 6.31s\nExit playbook, if user did not mean to reboot systems ------------------- 0.70s\nReboot system - wait for the reboot to complete ------------------------- 0.58s\n+ osism apply wait-for-connection -l testbed-nodes -e ireallymeanit=yes\n\u001b[32m2026-03-21 00:37:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of wait-for-connection.\u001b[0m\n\u001b[32m2026-03-21 00:37:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d0bc427e-2de9-41e8-a8db-b0f3e60a1221 (wait-for-connection) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 00:37:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task d0bc427e-2de9-41e8-a8db-b0f3e60a1221 (wait-for-connection) has been started and output is visible here.\u001b[0m\n\nPLAY [Wait until remote systems are reachable] *********************************\n\nTASK [Wait until remote system is reachable] ***********************************\nSaturday 21 March 2026  00:37:07 +0000 (0:00:00.310)       0:00:00.310 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:37:19 +0000 (0:00:11.510)       0:00:11.821 ******** \n=============================================================================== \nWait until remote system is reachable ---------------------------------- 11.51s\n+ osism apply hddtemp\n\u001b[32m2026-03-21 00:37:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of hddtemp.\u001b[0m\n\u001b[32m2026-03-21 00:37:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f636c49-3ee5-46f3-b998-ad491cd36695 (hddtemp) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 00:37:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 3f636c49-3ee5-46f3-b998-ad491cd36695 (hddtemp) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role hddtemp] ******************************************************\n\nTASK [osism.services.hddtemp : Gather variables for each operating system] *****\nSaturday 21 March 2026  00:37:33 +0000 (0:00:00.281)       0:00:00.281 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.hddtemp : Include distribution specific install tasks] ****\nSaturday 21 March 2026  00:37:34 +0000 (0:00:00.543)       0:00:00.825 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/hddtemp/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.hddtemp : Remove hddtemp package] *************************\nSaturday 21 March 2026  00:37:35 +0000 (0:00:01.039)       0:00:01.864 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.services.hddtemp : Enable Kernel Module drivetemp] *****************\nSaturday 21 March 2026  00:37:37 +0000 (0:00:02.460)       0:00:04.324 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.hddtemp : Check if drivetemp module is available] *********\nSaturday 21 March 2026  00:37:38 +0000 (0:00:01.047)       0:00:05.371 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.hddtemp : Load Kernel Module drivetemp] *******************\nSaturday 21 March 2026  00:37:40 +0000 (0:00:01.334)       0:00:06.705 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.hddtemp : Install lm-sensors] *****************************\nSaturday 21 March 2026  00:37:40 +0000 (0:00:00.607)       0:00:07.313 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.hddtemp : Include distribution specific service tasks] ****\nSaturday 21 March 2026  00:37:54 +0000 (0:00:13.421)       0:00:20.735 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/hddtemp/tasks/service-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.hddtemp : Manage lm-sensors service] **********************\nSaturday 21 March 2026  00:37:55 +0000 (0:00:01.050)       0:00:21.785 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:37:57 +0000 (0:00:02.460)       0:00:24.246 ******** \n=============================================================================== \nosism.services.hddtemp : Install lm-sensors ---------------------------- 13.42s\nosism.services.hddtemp : Manage lm-sensors service ---------------------- 2.46s\nosism.services.hddtemp : Remove hddtemp package ------------------------- 2.46s\nosism.services.hddtemp : Check if drivetemp module is available --------- 1.33s\nosism.services.hddtemp : Include distribution specific service tasks ---- 1.05s\nosism.services.hddtemp : Enable Kernel Module drivetemp ----------------- 1.05s\nosism.services.hddtemp : Include distribution specific install tasks ---- 1.04s\nosism.services.hddtemp : Load Kernel Module drivetemp ------------------- 0.61s\nosism.services.hddtemp : Gather variables for each operating system ----- 0.54s\n++ semver latest 7.1.1\n+ [[ -1 -ge 0 ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ sudo systemctl restart manager.service\n+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]\n+ wait_for_container_healthy 60 ceph-ansible\n+ local max_attempts=60\n+ local name=ceph-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ wait_for_container_healthy 60 kolla-ansible\n+ local max_attempts=60\n+ local name=kolla-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' kolla-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ wait_for_container_healthy 60 osism-ansible\n+ local max_attempts=60\n+ local name=osism-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' osism-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ [[ true == \\t\\r\\u\\e ]]\n+ sh -c /opt/configuration/scripts/disable-ara.sh\nARA in ceph-ansible already disabled.\nARA in kolla-ansible already disabled.\nARA in osism-ansible already disabled.\nARA in osism-kubernetes already disabled.\n+ osism apply gather-facts\n\u001b[32m2026-03-21 00:39:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of gather-facts.\u001b[0m\n\u001b[32m2026-03-21 00:39:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 25cfd3ec-85e1-4c1c-9913-f49fc78dae3a (gather-facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 00:39:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 25cfd3ec-85e1-4c1c-9913-f49fc78dae3a (gather-facts) has been started and output is visible here.\u001b[0m\n\u001b[1;35m[WARNING]: Invalid characters were found in group names but not replaced, use\u001b[0m\n\u001b[1;35m-vvvv to see details\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\n\u001b[1;31mfatal: [testbed-manager]: UNREACHABLE! => {\"changed\": false, \"msg\": \"Data could not be sent to remote host \\\"192.168.16.5\\\". Make sure this host can be reached over ssh: no such identity: /ansible/secrets/id_rsa: No such file or directory\\r\\ndragon@192.168.16.5: Permission denied (publickey).\\r\\n\", \"unreachable\": true}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\u001b[1;31mfatal: [testbed-node-5]: UNREACHABLE! => {\"changed\": false, \"msg\": \"Data could not be sent to remote host \\\"192.168.16.15\\\". Make sure this host can be reached over ssh: no such identity: /ansible/secrets/id_rsa: No such file or directory\\r\\ndragon@192.168.16.15: Permission denied (publickey).\\r\\n\", \"unreachable\": true}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\u001b[1;31mfatal: [testbed-node-1]: UNREACHABLE! => {\"changed\": false, \"msg\": \"Data could not be sent to remote host \\\"192.168.16.11\\\". Make sure this host can be reached over ssh: no such identity: /ansible/secrets/id_rsa: No such file or directory\\r\\ndragon@192.168.16.11: Permission denied (publickey).\\r\\n\", \"unreachable\": true}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\u001b[1;31mfatal: [testbed-node-2]: UNREACHABLE! => {\"changed\": false, \"msg\": \"Data could not be sent to remote host \\\"192.168.16.12\\\". Make sure this host can be reached over ssh: no such identity: /ansible/secrets/id_rsa: No such file or directory\\r\\ndragon@192.168.16.12: Permission denied (publickey).\\r\\n\", \"unreachable\": true}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\u001b[1;31mfatal: [testbed-node-3]: UNREACHABLE! => {\"changed\": false, \"msg\": \"Data could not be sent to remote host \\\"192.168.16.13\\\". Make sure this host can be reached over ssh: no such identity: /ansible/secrets/id_rsa: No such file or directory\\r\\ndragon@192.168.16.13: Permission denied (publickey).\\r\\n\", \"unreachable\": true}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\u001b[1;31mfatal: [testbed-node-0]: UNREACHABLE! => {\"changed\": false, \"msg\": \"Data could not be sent to remote host \\\"192.168.16.10\\\". Make sure this host can be reached over ssh: no such identity: /ansible/secrets/id_rsa: No such file or directory\\r\\ndragon@192.168.16.10: Permission denied (publickey).\\r\\n\", \"unreachable\": true}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\u001b[1;31mfatal: [testbed-node-4]: UNREACHABLE! => {\"changed\": false, \"msg\": \"Data could not be sent to remote host \\\"192.168.16.14\\\". Make sure this host can be reached over ssh: no such identity: /ansible/secrets/id_rsa: No such file or directory\\r\\ndragon@192.168.16.14: Permission denied (publickey).\\r\\n\", \"unreachable\": true}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\nPLAY [Gather facts for all hosts if using --limit] *****************************\n\nTASK [Gather facts for all hosts] **********************************************\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\n+ sudo ln -sf /opt/configuration/scripts/deploy/001-helpers.sh /usr/local/bin/deploy-helper\n+ sudo ln -sf /opt/configuration/scripts/deploy/100-ceph-with-ansible.sh /usr/local/bin/deploy-ceph-with-ansible\n+ sudo ln -sf /opt/configuration/scripts/deploy/100-ceph-with-rook.sh /usr/local/bin/deploy-ceph-with-rook\n+ sudo ln -sf /opt/configuration/scripts/deploy/200-infrastructure.sh /usr/local/bin/deploy-infrastructure\n+ sudo ln -sf /opt/configuration/scripts/deploy/300-openstack.sh /usr/local/bin/deploy-openstack\n+ sudo ln -sf /opt/configuration/scripts/deploy/320-openstack-minimal.sh /usr/local/bin/deploy-openstack-minimal\n+ sudo ln -sf /opt/configuration/scripts/deploy/400-monitoring.sh /usr/local/bin/deploy-monitoring\n+ sudo ln -sf /opt/configuration/scripts/deploy/500-kubernetes.sh /usr/local/bin/deploy-kubernetes\n+ sudo ln -sf /opt/configuration/scripts/deploy/510-clusterapi.sh /usr/local/bin/deploy-kubernetes-clusterapi\n+ sudo ln -sf /opt/configuration/scripts/upgrade-manager.sh /usr/local/bin/upgrade-manager\n+ sudo ln -sf /opt/configuration/scripts/upgrade/100-ceph-with-ansible.sh /usr/local/bin/upgrade-ceph-with-ansible\n+ sudo ln -sf /opt/configuration/scripts/upgrade/100-ceph-with-rook.sh /usr/local/bin/upgrade-ceph-with-rook\n+ sudo ln -sf /opt/configuration/scripts/upgrade/200-infrastructure.sh /usr/local/bin/upgrade-infrastructure\n+ sudo ln -sf /opt/configuration/scripts/upgrade/300-openstack.sh /usr/local/bin/upgrade-openstack\n+ sudo ln -sf /opt/configuration/scripts/upgrade/320-openstack-minimal.sh /usr/local/bin/upgrade-openstack-minimal\n+ sudo ln -sf /opt/configuration/scripts/upgrade/400-monitoring.sh /usr/local/bin/upgrade-monitoring\n+ sudo ln -sf /opt/configuration/scripts/upgrade/500-kubernetes.sh /usr/local/bin/upgrade-kubernetes\n+ sudo ln -sf /opt/configuration/scripts/upgrade/510-clusterapi.sh /usr/local/bin/upgrade-kubernetes-clusterapi\n+ sudo ln -sf /opt/configuration/scripts/bootstrap/300-openstack.sh /usr/local/bin/bootstrap-openstack\n+ sudo ln -sf /opt/configuration/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh /usr/local/bin/bootstrap-octavia\n+ sudo ln -sf /opt/configuration/scripts/bootstrap/302-openstack-k8s-clusterapi-images.sh /usr/local/bin/bootstrap-clusterapi\n+ sudo ln -sf /opt/configuration/scripts/disable-local-registry.sh /usr/local/bin/disable-local-registry\n+ sudo ln -sf /opt/configuration/scripts/pull-images.sh /usr/local/bin/pull-images\n+ [[ false == \\t\\r\\u\\e ]]",
                            "stdout_lines": [
                                "",
                                "# DEPLOY MANAGER",
                                "",
                                "+ set -e",
                                "+ echo",
                                "+ echo '# DEPLOY MANAGER'",
                                "+ echo",
                                "+ cat /opt/manager-vars.sh",
                                "export NUMBER_OF_NODES=6",
                                "",
                                "export CEPH_VERSION=reef",
                                "export CONFIGURATION_VERSION=main",
                                "export MANAGER_VERSION=latest",
                                "export OPENSTACK_VERSION=2024.2",
                                "",
                                "export ARA=false",
                                "export DEPLOY_MODE=manager",
                                "export TEMPEST=true",
                                "export IS_ZUUL=true",
                                "",
                                "export MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136",
                                "export EXTERNAL_API=false",
                                "",
                                "export IMAGE_USER=ubuntu",
                                "export IMAGE_NODE_USER=ubuntu",
                                "",
                                "export CEPH_STACK=ceph-ansible",
                                "",
                                "+ echo",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=true",
                                "++ TEMPEST=true",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ sudo ln -sf /opt/configuration/contrib/semver2.sh /usr/local/bin/semver",
                                "+ docker version",
                                "Client: Docker Engine - Community",
                                " Version:           27.5.1",
                                " API version:       1.47",
                                " Go version:        go1.22.11",
                                " Git commit:        9f9e405",
                                " Built:             Wed Jan 22 13:41:48 2025",
                                " OS/Arch:           linux/amd64",
                                " Context:           default",
                                "",
                                "Server: Docker Engine - Community",
                                " Engine:",
                                "  Version:          27.5.1",
                                "  API version:      1.47 (minimum version 1.24)",
                                "  Go version:       go1.22.11",
                                "  Git commit:       4c9b3b0",
                                "  Built:            Wed Jan 22 13:41:48 2025",
                                "  OS/Arch:          linux/amd64",
                                "  Experimental:     false",
                                " containerd:",
                                "  Version:          v2.2.2",
                                "  GitCommit:        301b2dac98f15c27117da5c8af12118a041a31d9",
                                " runc:",
                                "  Version:          1.3.4",
                                "  GitCommit:        v1.3.4-0-gd6d73eb8",
                                " docker-init:",
                                "  Version:          0.19.0",
                                "  GitCommit:        de40ad0",
                                "+ sh -c /opt/configuration/scripts/deploy/000-manager.sh",
                                "+ set -e",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=true",
                                "++ TEMPEST=true",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ /opt/configuration/scripts/set-ceph-version.sh reef",
                                "+ set -e",
                                "+ VERSION=reef",
                                "++ grep '^ceph_version:' /opt/configuration/environments/manager/configuration.yml",
                                "+ [[ -n ceph_version: reef ]]",
                                "+ sed -i 's/ceph_version: .*/ceph_version: reef/g' /opt/configuration/environments/manager/configuration.yml",
                                "+ /opt/configuration/scripts/set-openstack-version.sh 2024.2",
                                "+ set -e",
                                "+ VERSION=2024.2",
                                "++ grep '^openstack_version:' /opt/configuration/environments/manager/configuration.yml",
                                "+ [[ -n openstack_version: 2024.2 ]]",
                                "+ sed -i 's/openstack_version: .*/openstack_version: 2024.2/g' /opt/configuration/environments/manager/configuration.yml",
                                "+ [[ ceph-ansible == \\r\\o\\o\\k ]]",
                                "++ semver latest 7.0.0",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ echo 'enable_osism_kubernetes: true'",
                                "++ semver latest 10.0.0-0",
                                "+ [[ -1 -ge 0 ]]",
                                "++ semver 2024.2 2025.1",
                                "+ [[ -1 -ge 0 ]]",
                                "+ /opt/configuration/scripts/enable-resource-nodes.sh",
                                "+ [[ -e /opt/venv/bin/activate ]]",
                                "+ source /opt/venv/bin/activate",
                                "++ deactivate nondestructive",
                                "++ '[' -n '' ']'",
                                "++ '[' -n '' ']'",
                                "++ hash -r",
                                "++ '[' -n '' ']'",
                                "++ unset VIRTUAL_ENV",
                                "++ unset VIRTUAL_ENV_PROMPT",
                                "++ '[' '!' nondestructive = nondestructive ']'",
                                "++ '[' linux-gnu = cygwin ']'",
                                "++ '[' linux-gnu = msys ']'",
                                "++ export VIRTUAL_ENV=/opt/venv",
                                "++ VIRTUAL_ENV=/opt/venv",
                                "++ _OLD_VIRTUAL_PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin",
                                "++ PATH=/opt/venv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin",
                                "++ export PATH",
                                "++ '[' -n '' ']'",
                                "++ '[' -z '' ']'",
                                "++ _OLD_VIRTUAL_PS1=",
                                "++ PS1='(venv) '",
                                "++ export PS1",
                                "++ VIRTUAL_ENV_PROMPT='(venv) '",
                                "++ export VIRTUAL_ENV_PROMPT",
                                "++ hash -r",
                                "+ ansible-playbook -i testbed-manager, --vault-password-file /opt/configuration/environments/.vault_pass /opt/configuration/ansible/manager-part-3.yml",
                                "",
                                "PLAY [Copy custom facts] *******************************************************",
                                "",
                                "TASK [Create custom facts directory] *******************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Copy fact files] *********************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "PLAY [Before the deployment of the manager] ************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Get /opt/manager-vars.sh] ************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Add ara_server_mariadb_volume_type parameter] ****************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Add netbox_enable parameter] *********************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Install HWE kernel package on Ubuntu] ************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Check if /etc/OTC_region exist] ******************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Add nova_compute_virt_type parameter] ************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "PLAY [Apply role traefik] ******************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Apply traefik role] ******************************************************",
                                "included: osism.services.traefik for testbed-manager",
                                "",
                                "TASK [osism.services.traefik : Include config tasks] ***************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/traefik/tasks/config.yml for testbed-manager",
                                "",
                                "TASK [osism.services.traefik : Create required directories] ********************",
                                "changed: [testbed-manager] => (item=/opt/traefik)",
                                "changed: [testbed-manager] => (item=/opt/traefik/certificates)",
                                "changed: [testbed-manager] => (item=/opt/traefik/configuration)",
                                "",
                                "TASK [osism.services.traefik : Copy configuration files] ***********************",
                                "changed: [testbed-manager] => (item=traefik.yml)",
                                "changed: [testbed-manager] => (item=traefik.env)",
                                "changed: [testbed-manager] => (item=certificates.yml)",
                                "",
                                "TASK [osism.services.traefik : Copy certificate cert files] ********************",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Copy certificate key files] *********************",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Copy dynamic configuration] *********************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Remove dynamic configuration] *******************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Include service tasks] **************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/traefik/tasks/service.yml for testbed-manager",
                                "",
                                "TASK [osism.services.traefik : Create traefik external network] ****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Copy docker-compose.yml file] *******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Manage traefik service] *************************",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.traefik : Restart traefik service] *************",
                                "skipping: [testbed-manager]",
                                "",
                                "PLAY [Deploy manager service] **************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Apply manager role] ******************************************************",
                                "included: osism.services.manager for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Include install tasks] **************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/install-Debian-family.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Install required packages] **********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Gather variables for each operating system] *****",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include config tasks] ***************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Create required directories] ********************",
                                "changed: [testbed-manager] => (item=/opt/ansible)",
                                "changed: [testbed-manager] => (item=/opt/archive)",
                                "changed: [testbed-manager] => (item=/opt/manager/configuration)",
                                "changed: [testbed-manager] => (item=/opt/manager/data)",
                                "ok: [testbed-manager] => (item=/opt/manager)",
                                "changed: [testbed-manager] => (item=/opt/manager/secrets)",
                                "changed: [testbed-manager] => (item=/opt/ansible/secrets)",
                                "changed: [testbed-manager] => (item=/opt/state)",
                                "",
                                "TASK [osism.services.manager : Copy all environment file] **********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy client environment file] *******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include ara config tasks] ***********************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ara.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Copy ARA environment files] *********************",
                                "changed: [testbed-manager] => (item=ara)",
                                "changed: [testbed-manager] => (item=ara-server)",
                                "",
                                "TASK [osism.services.manager : Copy MariaDB environment file] ******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include vault config tasks] *********************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include frontend config tasks] ******************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-frontend.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Copy frontend environment file] *****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include ansible config tasks] *******************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ansible.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Copy private ssh keys] **************************",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy ansible environment file] ******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include netbox config tasks] ********************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include celery config tasks] ********************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-celery.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Set fs.inotify.max_user_watches] ****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Set fs.inotify.max_user_instances] **************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy celery environment files] ******************",
                                "changed: [testbed-manager] => (item=conductor)",
                                "changed: [testbed-manager] => (item=openstack)",
                                "",
                                "TASK [osism.services.manager : Copy listener environment file] *****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Check for conductor.yml] ************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy conductor configuration file] **************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy empty conductor configuration file] ********",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include wrapper config tasks] *******************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-wrapper.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Include wrapper vars file] **********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy wrapper scripts] ***************************",
                                "changed: [testbed-manager] => (item=osism)",
                                "changed: [testbed-manager] => (item=osism-update-docker)",
                                "changed: [testbed-manager] => (item=osism-update-manager)",
                                "",
                                "TASK [osism.services.manager : Copy cilium wrapper script] *********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy hubble wrapper script] *********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy flux wrapper script] ***********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include scripts config tasks] *******************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-scripts.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Include scripts vars file] **********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy scripts] ***********************************",
                                "changed: [testbed-manager] => (item=osism-include)",
                                "",
                                "TASK [osism.services.manager : Include service tasks] **************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/service.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Copy manager systemd unit file] *****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Create traefik external network] ****************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Set mariadb healthcheck for mariadb < 11.0.0] ***",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Set mariadb healthcheck for mariadb >= 11.0.0] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy docker-compose.yml file] *******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Pull container images] **************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Stop and disable old service docker-compose@manager] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Do a manual start of the manager service] *******",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Manage manager service] *************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Register that manager service was started] ******",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Flush handlers] *********************************",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Restart manager service] *************",
                                "skipping: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Wait for manager service to start] ***",
                                "Pausing for 60 seconds",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Ensure that all containers are up] ***",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Wait for an healthy manager service] ***",
                                "FAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (50 retries left).",
                                "FAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (49 retries left).",
                                "FAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (48 retries left).",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Copy osismclient bash completion script] ***",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include initialize tasks] ***********************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/initialize.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Flush handlers] *********************************",
                                "",
                                "TASK [osism.services.manager : Include vault initialize tasks] *****************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include version verification tasks] *************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/verify-versions.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Deploy service manager version check script] ****",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Execute service manager version check] **********",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Display version check results] ******************",
                                "ok: [testbed-manager] => {",
                                "    \"version_check_result.stdout_lines\": [",
                                "        \"=== OSISM Container Version Check ===\",",
                                "        \"Checking running containers against expected versions...\",",
                                "        \"\",",
                                "        \"Checking service: inventory_reconciler (Inventory Reconciler Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/inventory-reconciler:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/inventory-reconciler:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: osism-ansible (OSISM Ansible Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism-ansible:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism-ansible:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: osism-kubernetes (Osism-Kubernetes Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism-kubernetes:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism-kubernetes:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: ceph-ansible (Ceph-Ansible Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/ceph-ansible:reef\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/ceph-ansible:reef\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: kolla-ansible (Kolla-Ansible Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/kolla-ansible:2024.2\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/kolla-ansible:2024.2\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: osismclient (OSISM Client)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: ara-server (ARA Server)\",",
                                "        \"  Expected: registry.osism.tech/osism/ara-server:1.7.3\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/ara-server:1.7.3\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: mariadb (MariaDB for ARA)\",",
                                "        \"  Expected: registry.osism.tech/dockerhub/library/mariadb:11.8.4\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/dockerhub/library/mariadb:11.8.4\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: frontend (OSISM Frontend)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism-frontend:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism-frontend:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: redis (Redis Cache)\",",
                                "        \"  Expected: registry.osism.tech/dockerhub/library/redis:7.4.7-alpine\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/dockerhub/library/redis:7.4.7-alpine\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: api (OSISM API Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: listener (OpenStack Event Listener)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: openstack (OpenStack Integration)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: beat (Celery Beat Scheduler)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: flower (Celery Flower Monitor)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"=== Summary ===\",",
                                "        \"Errors (version mismatches): 0\",",
                                "        \"Warnings (expected containers not running): 0\",",
                                "        \"\",",
                                "        \"\u2705 All running containers match expected versions!\"",
                                "    ]",
                                "}",
                                "",
                                "TASK [osism.services.manager : Skip version check due to service configuration] ***",
                                "skipping: [testbed-manager]",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "testbed-manager            : ok=70   changed=37   unreachable=0    failed=0    skipped=12   rescued=0    ignored=0   ",
                                "",
                                "+ [[ -e /opt/venv/bin/activate ]]",
                                "+ deactivate",
                                "+ '[' -n /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin ']'",
                                "+ PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin",
                                "+ export PATH",
                                "+ unset _OLD_VIRTUAL_PATH",
                                "+ '[' -n '' ']'",
                                "+ hash -r",
                                "+ '[' -n '' ']'",
                                "+ unset VIRTUAL_ENV",
                                "+ unset VIRTUAL_ENV_PROMPT",
                                "+ '[' '!' '' = nondestructive ']'",
                                "+ unset -f deactivate",
                                "+ cp /home/dragon/.ssh/id_rsa.pub /opt/ansible/secrets/id_rsa.operator.pub",
                                "+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]",
                                "+ wait_for_container_healthy 60 ceph-ansible",
                                "+ local max_attempts=60",
                                "+ local name=ceph-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ wait_for_container_healthy 60 kolla-ansible",
                                "+ local max_attempts=60",
                                "+ local name=kolla-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' kolla-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ wait_for_container_healthy 60 osism-ansible",
                                "+ local max_attempts=60",
                                "+ local name=osism-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' osism-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ [[ true == \\t\\r\\u\\e ]]",
                                "+ sh -c /opt/configuration/scripts/disable-ara.sh",
                                "+ docker compose --project-directory /opt/manager ps",
                                "NAME                             IMAGE                                                      COMMAND                  SERVICE                CREATED         STATUS                        PORTS",
                                "ceph-ansible                     registry.osism.tech/osism/ceph-ansible:reef                \"/entrypoint.sh osis\u2026\"   ceph-ansible           2 minutes ago   Up About a minute (healthy)   ",
                                "kolla-ansible                    registry.osism.tech/osism/kolla-ansible:2024.2             \"/entrypoint.sh osis\u2026\"   kolla-ansible          2 minutes ago   Up About a minute (healthy)   ",
                                "manager-api-1                    registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   api                    2 minutes ago   Up 2 minutes (healthy)        192.168.16.5:8000->8000/tcp",
                                "manager-ara-server-1             registry.osism.tech/osism/ara-server:1.7.3                 \"sh -c '/wait && /ru\u2026\"   ara-server             2 minutes ago   Up 2 minutes (healthy)        8000/tcp",
                                "manager-beat-1                   registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   beat                   2 minutes ago   Up 2 minutes (healthy)        ",
                                "manager-flower-1                 registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   flower                 2 minutes ago   Up 2 minutes (healthy)        ",
                                "manager-inventory_reconciler-1   registry.osism.tech/osism/inventory-reconciler:latest      \"/sbin/tini -- /entr\u2026\"   inventory_reconciler   2 minutes ago   Up About a minute (healthy)   ",
                                "manager-listener-1               registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   listener               2 minutes ago   Up 2 minutes (healthy)        ",
                                "manager-mariadb-1                registry.osism.tech/dockerhub/library/mariadb:11.8.4       \"docker-entrypoint.s\u2026\"   mariadb                2 minutes ago   Up 2 minutes (healthy)        3306/tcp",
                                "manager-openstack-1              registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   openstack              2 minutes ago   Up 2 minutes (healthy)        ",
                                "manager-redis-1                  registry.osism.tech/dockerhub/library/redis:7.4.7-alpine   \"docker-entrypoint.s\u2026\"   redis                  2 minutes ago   Up 2 minutes (healthy)        6379/tcp",
                                "osism-ansible                    registry.osism.tech/osism/osism-ansible:latest             \"/entrypoint.sh osis\u2026\"   osism-ansible          2 minutes ago   Up About a minute (healthy)   ",
                                "osism-frontend                   registry.osism.tech/osism/osism-frontend:latest            \"docker-entrypoint.s\u2026\"   frontend               2 minutes ago   Up 2 minutes                  192.168.16.5:3000->3000/tcp",
                                "osism-kubernetes                 registry.osism.tech/osism/osism-kubernetes:latest          \"/entrypoint.sh osis\u2026\"   osism-kubernetes       2 minutes ago   Up About a minute (healthy)   ",
                                "osismclient                      registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- sleep\u2026\"   osismclient            2 minutes ago   Up 2 minutes (healthy)        ",
                                "++ semver latest 7.0.0",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ sed -i s/community.general.yaml/osism.commons.still_alive/ /opt/configuration/environments/ansible.cfg",
                                "+ osism apply resolvconf -l testbed-manager",
                                "\u001b[32m2026-03-21 00:20:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of resolvconf.\u001b[0m",
                                "\u001b[32m2026-03-21 00:20:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1cd44190-d184-4b83-bd15-79b0234c9cb3 (resolvconf) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 00:20:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 1cd44190-d184-4b83-bd15-79b0234c9cb3 (resolvconf) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role resolvconf] ***************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Saturday 21 March 2026  00:20:36 +0000 (0:00:00.198)       0:00:00.198 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Check minimum and maximum number of name servers] ***",
                                "Saturday 21 March 2026  00:20:41 +0000 (0:00:04.929)       0:00:05.127 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include resolvconf tasks] *********************",
                                "Saturday 21 March 2026  00:20:41 +0000 (0:00:00.068)       0:00:05.196 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-resolv.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include distribution specific installation tasks] ***",
                                "Saturday 21 March 2026  00:20:41 +0000 (0:00:00.093)       0:00:05.290 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf] ***",
                                "Saturday 21 March 2026  00:20:41 +0000 (0:00:00.079)       0:00:05.370 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Install package systemd-resolved] *************",
                                "Saturday 21 March 2026  00:20:42 +0000 (0:00:01.120)       0:00:06.490 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf] *****",
                                "Saturday 21 March 2026  00:20:43 +0000 (0:00:00.056)       0:00:06.546 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Archive existing file /etc/resolv.conf] *******",
                                "Saturday 21 March 2026  00:20:43 +0000 (0:00:00.509)       0:00:07.056 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf] ***",
                                "Saturday 21 March 2026  00:20:43 +0000 (0:00:00.074)       0:00:07.131 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Copy configuration files] *********************",
                                "Saturday 21 March 2026  00:20:44 +0000 (0:00:00.533)       0:00:07.665 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Start/enable systemd-resolved service] ********",
                                "Saturday 21 March 2026  00:20:45 +0000 (0:00:01.054)       0:00:08.719 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include distribution specific configuration tasks] ***",
                                "Saturday 21 March 2026  00:20:46 +0000 (0:00:00.920)       0:00:09.640 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Restart systemd-resolved service] *************",
                                "Saturday 21 March 2026  00:20:46 +0000 (0:00:00.083)       0:00:09.724 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:20:47 +0000 (0:00:01.106)       0:00:10.831 ******** ",
                                "=============================================================================== ",
                                "Gathering Facts --------------------------------------------------------- 4.93s",
                                "osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf --- 1.12s",
                                "osism.commons.resolvconf : Restart systemd-resolved service ------------- 1.11s",
                                "osism.commons.resolvconf : Copy configuration files --------------------- 1.05s",
                                "osism.commons.resolvconf : Start/enable systemd-resolved service -------- 0.92s",
                                "osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf --- 0.53s",
                                "osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf ----- 0.51s",
                                "osism.commons.resolvconf : Include resolvconf tasks --------------------- 0.09s",
                                "osism.commons.resolvconf : Include distribution specific configuration tasks --- 0.08s",
                                "osism.commons.resolvconf : Include distribution specific installation tasks --- 0.08s",
                                "osism.commons.resolvconf : Archive existing file /etc/resolv.conf ------- 0.07s",
                                "osism.commons.resolvconf : Check minimum and maximum number of name servers --- 0.07s",
                                "osism.commons.resolvconf : Install package systemd-resolved ------------- 0.06s",
                                "+ osism apply sshconfig",
                                "\u001b[32m2026-03-21 00:20:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of sshconfig.\u001b[0m",
                                "\u001b[32m2026-03-21 00:20:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask edf4ac79-9196-4576-ba98-18766b3fa2b1 (sshconfig) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 00:20:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task edf4ac79-9196-4576-ba98-18766b3fa2b1 (sshconfig) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role sshconfig] ****************************************************",
                                "",
                                "TASK [osism.commons.sshconfig : Get home directory of operator user] ***********",
                                "Saturday 21 March 2026  00:21:01 +0000 (0:00:00.144)       0:00:00.144 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.sshconfig : Ensure .ssh/config.d exist] ********************",
                                "Saturday 21 March 2026  00:21:02 +0000 (0:00:00.859)       0:00:01.004 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.sshconfig : Ensure config for each host exist] *************",
                                "Saturday 21 March 2026  00:21:03 +0000 (0:00:00.474)       0:00:01.479 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5)\u001b[0m",
                                "",
                                "TASK [osism.commons.sshconfig : Add extra config] ******************************",
                                "Saturday 21 March 2026  00:21:08 +0000 (0:00:05.164)       0:00:06.643 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.sshconfig : Assemble ssh config] ***************************",
                                "Saturday 21 March 2026  00:21:08 +0000 (0:00:00.108)       0:00:06.752 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=4   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:21:09 +0000 (0:00:00.516)       0:00:07.268 ******** ",
                                "=============================================================================== ",
                                "osism.commons.sshconfig : Ensure config for each host exist ------------- 5.16s",
                                "osism.commons.sshconfig : Get home directory of operator user ----------- 0.86s",
                                "osism.commons.sshconfig : Assemble ssh config --------------------------- 0.52s",
                                "osism.commons.sshconfig : Ensure .ssh/config.d exist -------------------- 0.47s",
                                "osism.commons.sshconfig : Add extra config ------------------------------ 0.11s",
                                "+ osism apply known-hosts",
                                "\u001b[32m2026-03-21 00:21:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of known-hosts.\u001b[0m",
                                "\u001b[32m2026-03-21 00:21:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c84cdb3b-fb8b-4d97-9693-70eeb7a327d5 (known-hosts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 00:21:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task c84cdb3b-fb8b-4d97-9693-70eeb7a327d5 (known-hosts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role known_hosts] **************************************************",
                                "",
                                "TASK [osism.commons.known_hosts : Run ssh-keyscan for all hosts with hostname] ***",
                                "Saturday 21 March 2026  00:21:23 +0000 (0:00:00.193)       0:00:00.193 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-manager)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-5)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with hostname] ***",
                                "Saturday 21 March 2026  00:21:30 +0000 (0:00:06.436)       0:00:06.629 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-manager)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-0)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-1)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-2)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-3)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-4)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-5)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Saturday 21 March 2026  00:21:30 +0000 (0:00:00.192)       0:00:06.821 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBGCW8S4XhFwtxakQShgXgO7SwC1ojlJgYP5uttdiBrf1czDr4b2JfGHDaBJkhAR0HWZVp4aSpl8XO7vpRuLHcKg=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCRlOzWKiVjhf3/9gUxrgBZmRlENj/bm0EeQnimGnkItBo13vP/sIQySfQ4wuvVN2E7kjt0LDhutlgBP+9+9QmN7QyxACxIVy+Yv3TDwZaTwkeU71pgeZ2nuphSGxEwWxtpOk6NTrdk8LCUuXK5IzKa++u+/ESIaPuIjMUQpLsc7m6taSCDlQwhMPbh4pkUCOfTKvWTZvUTzhfXd0v2yu2oYE6HKi552UF1xZeFUdWj2mLvuFVtObaVPiqDrfl80u0mOYPZzkeHGg4OwJqr8+eYNrTy2PPcvpPJz2zd6KCKrScr0YqlFrI1/iqMEdB24un2H9hiGm21ETl7jBPeYUSkY8fwR/z18SGIEqcwxLwdyEVufwpclNnJpLdeueDqOTUWSzm/2eV9xxQRUHv37+cdQHGlsw9Vkd1vB7S+RyiMCxHvzM+F7s9qheuk1oN1YBk0Gu9PirJddsCb13/vR9Uq3aGR8Rl+JNQjBLZ52WoiBOU6x9Xwl06LIMXvTYbk6R8=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIINhMRJZFRnFvt0mw99z3/Oe5sZY19PiSKPeN08h+FZs)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Saturday 21 March 2026  00:21:31 +0000 (0:00:01.224)       0:00:08.046 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDBFYMv4DFwxq5kkUW4Yk1dzqvK6rxV/EnP+KlAQYoqgq0Jxy6RiZBjZYrs44Cz5N22qgQKRJOIKZeviB2HMPXP9M/stCMvrwKHQ759UIXdlCFSM1wb/XiWVEfpvyg17SK8C6Sk3TFQNxDBvH8tuoB9itW69HmiDzMNQzJjw+8rbaAFfrJVxweUkVfjVL3u1XXMaTcarhWQ8AM1qCt8WSpc87HZ+5nPz6MiwzvnlbXpGvmPDH8jpVGGI0liY0uuXbc+cVSyVHpHqcS+0btiGc2o8CrV86oBz1kBA2QlztckLU8o0J9EsO9tXNBFf9dXpgLoq/3hsyFh5J3zWQK6FlyaRX0fiwTKLyPuoeaIn8rjM6YLfZCdJW+dFfo7mmN4FJaLcQz0flwy5rGNztyjK76Zij71zy4a5YNJoqnoHlOkp12AwJZVX+UrXTpPircM/LObMS26OBZousjzYSsowqdR1qwHJkth5DI9wqPdIj0OYb8w422VthxdB7jWvJDL+yM=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBHh3FDIOKYfKw5RWzPvfahw1uwX3EnFPFc7j151CAfmuNXNhWNd/g7+a+BcFHxWoUVLQ2bHNkBvDQ853xelpFss=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIJtBRNCPeE2jW0fyck0ZOiM7FLqI8afu36mE896gdALV)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Saturday 21 March 2026  00:21:32 +0000 (0:00:01.021)       0:00:09.068 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCkOmCQltaAtHo3qnWk8Zva2/YySzFSuTk33gJ8qFQSOoJ/JMivrQOeeT/xhyTpT0KdDa4wUEJ2DFsRRJtQ4RJT1tOvvKVcOHCUAb5TukWWcRLzojdpJje3FzjBhm3CIYeeCVmUFQob4+Fbsvq14fYAJwJ+iLRVb3WYjJ2DQpdgnSizr+YXC1FgaUAHsI0qKtZXRFHs3kRz/V086+0QhcVvLEVkzCcrUJpK3br38IxBNtUt4riGvSJeBAEZjPDCoZYULNKhfuDJpSTXtFhDPURIdVGMiphFAyJc75UXDmAur9PgARiKpIUiW6lkK99pEzQL81mDPGgfR2IVt6WyHB6+KHsyGBkAhV9Hl/mu0wPs/BQJIiwzKKEZwjwQHZtH2WbPY5ZO01+mbR2B/mdJIyIBPOT70tA3oF/i1dOm2akpHMA5b6Ygn6LgNXZ2/R/ONzk5FqfWdjAyR/g8trdF5OeFSbs8+yZjmixmGFlXUgzP09SPhzcaZr6eu6Re8ubRUCk=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPidv0r0leu7YKpNvc3qywT5qo1UxOebZ4/1THmXIT8p95UG8xxtEgxjB6LiDlrIKHnOQcrjX0aPZd9tK0OUMH0=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIBXGRQ9OqD5v29iwd1ZjIplLtCmKlBxfmQ/Z4YtqKcvG)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Saturday 21 March 2026  00:21:33 +0000 (0:00:01.046)       0:00:10.114 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDLUD36gZZ07RR35hp8bzvE4EMQSFjQPExYfzrUx2UDw1Iw7hZO7DIVqdOYeO71sHdJCALdChBOd+sqrPhZVHU2pte0NRFLiKlfHPGtT8En+gvydegKX9ZrbDxH/QDQzAoK3oxa6rSY/hgrjalxSfFsoX3ck0+B+BAsnPQkl7pJb3qMxKbwj0BZg8hvl3ORoU2LywtrmZCwnldNdexRDw9WuG4A7CUtBAw0sRk1F+6XMVDBpgPy7Z/VKL/77GNdzZhfHnWRoZR16U6OiM4km40O9HhGpfguJXF7QS5O+I1qOeLnqkEY8C0Iw6VfLDril4+55vdrRO/301+/2sR34VdUinHwzvz8CK/gfAQFfOHQKtD5DbfNdAYJjchZmSd5QsQcJNCk+Oi3zJ+W30kkTnRWfK+X7O4YZ1Vr14gfuhEpWImGUVEKZ1MqIcf6qoCKGlAbz4Ftv6g9CAy+X2LQgMrYK7AsSPDFwQ2IXLkpNB8885jAXlajrYNFi+BuHilCJjM=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBKVwUgL19RWciqX6jXMbqinVkklJq2z1AvXg256NiASkiG1Mj5ak7uFZEkNGaEHf+K1QBoUh7fbzGpgkRUarIPU=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDo385EdO3Q9a9gM7vB54U7tO40OJjhIKuCD/4Lczktc)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Saturday 21 March 2026  00:21:34 +0000 (0:00:01.037)       0:00:11.151 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBMw16gD3Pbm/obmA1y0WkDRYOE6nz7IznqcYxNJs1Ttz0TfWx4+l2TZZB+cWEZlqH1wXHeH4gIujSHjYb0z47x4=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCzB2yNwrdxiybJKIFTUFKw9hgA3eTgnpnh7Bnn2LQx2kmV9Tb3SUpq4hu3HGINP8vFI02QBpLJY3J8calXDdlmDA7s+90KFx+ituDeWizWVTuBtnDfWKH+bXqnpK4fytWz+rraBQg0zENXnkq95/sFYQAlEfklsEN6QtaTqx9UBFXWyST4nveGW9WLbfvqaBiGw67W1alnRV39FUhctaQRVoxfGLmERo7vX7PAhq+olypDXKriYYrJUJYej+TBdEhBQKbjHxl+ltGYYsCAgQ3YNMicYsutcKJq1fn5y1xhkGh2D/3jMSj8zE3H/ZbwyEuwMvRce6WxPHyfRwwEEcw1z+ZWS2fr565/aK/d+oa+dPgmHdAkOHTLTTw5GHE32YXuhZ+oqUORH5NyTeTtl0jj+JFCUt1rEN+4ZyNQQtIX4k4bxce/K7hFeAmBE3oxWq8RaaR95g89U5m1CmJ22hK1sCZ6+4kHW4gmhwZU0tVuv+dGZHtzxNdHQhnjrFnkEfk=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIPChxonu3Hcdz1NXmezrajxuZTmJk01UE7s701krwJ8B)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Saturday 21 March 2026  00:21:35 +0000 (0:00:01.037)       0:00:12.189 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLaFmk9jZnhiVtKL0wRcGE4qVsljEsur5kTWwGDIJ1BGW9608838oko7fKNiNcyELZFzn+EJxiyJc06R/YAnm94=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAILcvUHyiPwzNQOC7IOncTN/c7TkgoFEUaxcoF1QveIN9)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC+2o4j0grHvLW3HhVtn4eHEakT9YYI3Sb++Ci0amagE5qNEbs/ZxyR9lzC65lYaEb3k10xr1I7dsrCB4xGyg89vnvU9BhFTChCpOBN4XSsQI0VXJcoK1gwh1vVcAgzoYbDh9qj25CpXEKmPS79tu5aMhzf7ENIjYP84/DyI7o/3zRk3GIocva42Ojjj8/EmmhNU0q+tfq4+kHnI6xCSoDK9YINST/sakm5VZWgzvZIluw7dbpbOvCgfflUmbaoO2emen9a1T23P/2RQLaCnVoiOij5+TIM4yZCPtFmg8Pf+BgRPt7dkjngX62KzF5NbXSNQGD7jI7L2bBWaOdKx1SmxbjIbQ/+ZwOcaUtfC49XfHHRrCQv6WXziDCPOSe+11xrCaHKcJMEztUEjJAR1Q0PI/bShfT+8TsULIMm1PYdKL+97cN57ccRrW2BEh1UExjG1LpvT+F0ihjouDQtOJdQTvGJHfAIwk0ttixM13ny9r/YvhZcZ8+OHpx75ljTc20=)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Saturday 21 March 2026  00:21:36 +0000 (0:00:01.039)       0:00:13.228 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBJcLJ530VGI+fZqYZivjtRjY60XsOdACkcMGZCO6DM1QCtAMis3Vt7RkgrBST8pYU1j4GVKr/b830ZCxBx8OWD8=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC8TUaJpxO/zwNXdrLifgopr+lsK/T9T7LNe2Pa4nLlZD69HGd8aTTUK8isoyIT+NnnT1Hna9SMLxNv8ibVyw6P7m3g8I6jzlc+J5ex443F1eIuwKlHiLKGrCCX+zp7DFbbJhcXDfl8cglHXPgdeyGVbDEWqOmWRSalh7QWAdF4dhVKp2J+rmuqJEz3MxqjNfpelItlm8Cd9dAL22kjI8YeqXTWW/q/5bkEwvApD5PqoxOe8FpSPtQUW1xx9/pwWyIbsN6ARcX8nrmzAHbwpHvc0GR39x1ahcLnFm+7UE+5Mxwq8SvCTQe2oeLkEg7XRdj3nJ7cUmGFTGczVMHULU6ZuHiabMhbwVbBBnjm5mfkCcOVdMOaYEAcqUBB51w/6RQpB+I23Di40MibLigFEoQOaIEHZtggAivOaXouqPlNQ21MevgBw1jzwVNtR44U4uBS2uMSrCeopIuxVYptnlOFa3YdDeS/gfAacyD3/dkLymC0rC4QBn1PjV46J6bbQ20=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIIdSvMWwXHWpnHDy//kDiEa/cHDwBq24RvQjo19fNzjB)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Run ssh-keyscan for all hosts with ansible_host] ***",
                                "Saturday 21 March 2026  00:21:38 +0000 (0:00:01.072)       0:00:14.300 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-manager)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-5)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with ansible_host] ***",
                                "Saturday 21 March 2026  00:21:43 +0000 (0:00:05.317)       0:00:19.618 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-manager)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-0)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-1)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-2)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-3)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-4)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-5)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Saturday 21 March 2026  00:21:43 +0000 (0:00:00.169)       0:00:19.787 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCRlOzWKiVjhf3/9gUxrgBZmRlENj/bm0EeQnimGnkItBo13vP/sIQySfQ4wuvVN2E7kjt0LDhutlgBP+9+9QmN7QyxACxIVy+Yv3TDwZaTwkeU71pgeZ2nuphSGxEwWxtpOk6NTrdk8LCUuXK5IzKa++u+/ESIaPuIjMUQpLsc7m6taSCDlQwhMPbh4pkUCOfTKvWTZvUTzhfXd0v2yu2oYE6HKi552UF1xZeFUdWj2mLvuFVtObaVPiqDrfl80u0mOYPZzkeHGg4OwJqr8+eYNrTy2PPcvpPJz2zd6KCKrScr0YqlFrI1/iqMEdB24un2H9hiGm21ETl7jBPeYUSkY8fwR/z18SGIEqcwxLwdyEVufwpclNnJpLdeueDqOTUWSzm/2eV9xxQRUHv37+cdQHGlsw9Vkd1vB7S+RyiMCxHvzM+F7s9qheuk1oN1YBk0Gu9PirJddsCb13/vR9Uq3aGR8Rl+JNQjBLZ52WoiBOU6x9Xwl06LIMXvTYbk6R8=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBGCW8S4XhFwtxakQShgXgO7SwC1ojlJgYP5uttdiBrf1czDr4b2JfGHDaBJkhAR0HWZVp4aSpl8XO7vpRuLHcKg=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIINhMRJZFRnFvt0mw99z3/Oe5sZY19PiSKPeN08h+FZs)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Saturday 21 March 2026  00:21:44 +0000 (0:00:01.025)       0:00:20.812 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDBFYMv4DFwxq5kkUW4Yk1dzqvK6rxV/EnP+KlAQYoqgq0Jxy6RiZBjZYrs44Cz5N22qgQKRJOIKZeviB2HMPXP9M/stCMvrwKHQ759UIXdlCFSM1wb/XiWVEfpvyg17SK8C6Sk3TFQNxDBvH8tuoB9itW69HmiDzMNQzJjw+8rbaAFfrJVxweUkVfjVL3u1XXMaTcarhWQ8AM1qCt8WSpc87HZ+5nPz6MiwzvnlbXpGvmPDH8jpVGGI0liY0uuXbc+cVSyVHpHqcS+0btiGc2o8CrV86oBz1kBA2QlztckLU8o0J9EsO9tXNBFf9dXpgLoq/3hsyFh5J3zWQK6FlyaRX0fiwTKLyPuoeaIn8rjM6YLfZCdJW+dFfo7mmN4FJaLcQz0flwy5rGNztyjK76Zij71zy4a5YNJoqnoHlOkp12AwJZVX+UrXTpPircM/LObMS26OBZousjzYSsowqdR1qwHJkth5DI9wqPdIj0OYb8w422VthxdB7jWvJDL+yM=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBHh3FDIOKYfKw5RWzPvfahw1uwX3EnFPFc7j151CAfmuNXNhWNd/g7+a+BcFHxWoUVLQ2bHNkBvDQ853xelpFss=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIJtBRNCPeE2jW0fyck0ZOiM7FLqI8afu36mE896gdALV)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Saturday 21 March 2026  00:21:45 +0000 (0:00:01.034)       0:00:21.847 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIBXGRQ9OqD5v29iwd1ZjIplLtCmKlBxfmQ/Z4YtqKcvG)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCkOmCQltaAtHo3qnWk8Zva2/YySzFSuTk33gJ8qFQSOoJ/JMivrQOeeT/xhyTpT0KdDa4wUEJ2DFsRRJtQ4RJT1tOvvKVcOHCUAb5TukWWcRLzojdpJje3FzjBhm3CIYeeCVmUFQob4+Fbsvq14fYAJwJ+iLRVb3WYjJ2DQpdgnSizr+YXC1FgaUAHsI0qKtZXRFHs3kRz/V086+0QhcVvLEVkzCcrUJpK3br38IxBNtUt4riGvSJeBAEZjPDCoZYULNKhfuDJpSTXtFhDPURIdVGMiphFAyJc75UXDmAur9PgARiKpIUiW6lkK99pEzQL81mDPGgfR2IVt6WyHB6+KHsyGBkAhV9Hl/mu0wPs/BQJIiwzKKEZwjwQHZtH2WbPY5ZO01+mbR2B/mdJIyIBPOT70tA3oF/i1dOm2akpHMA5b6Ygn6LgNXZ2/R/ONzk5FqfWdjAyR/g8trdF5OeFSbs8+yZjmixmGFlXUgzP09SPhzcaZr6eu6Re8ubRUCk=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPidv0r0leu7YKpNvc3qywT5qo1UxOebZ4/1THmXIT8p95UG8xxtEgxjB6LiDlrIKHnOQcrjX0aPZd9tK0OUMH0=)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Saturday 21 March 2026  00:21:46 +0000 (0:00:01.021)       0:00:22.869 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBKVwUgL19RWciqX6jXMbqinVkklJq2z1AvXg256NiASkiG1Mj5ak7uFZEkNGaEHf+K1QBoUh7fbzGpgkRUarIPU=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDLUD36gZZ07RR35hp8bzvE4EMQSFjQPExYfzrUx2UDw1Iw7hZO7DIVqdOYeO71sHdJCALdChBOd+sqrPhZVHU2pte0NRFLiKlfHPGtT8En+gvydegKX9ZrbDxH/QDQzAoK3oxa6rSY/hgrjalxSfFsoX3ck0+B+BAsnPQkl7pJb3qMxKbwj0BZg8hvl3ORoU2LywtrmZCwnldNdexRDw9WuG4A7CUtBAw0sRk1F+6XMVDBpgPy7Z/VKL/77GNdzZhfHnWRoZR16U6OiM4km40O9HhGpfguJXF7QS5O+I1qOeLnqkEY8C0Iw6VfLDril4+55vdrRO/301+/2sR34VdUinHwzvz8CK/gfAQFfOHQKtD5DbfNdAYJjchZmSd5QsQcJNCk+Oi3zJ+W30kkTnRWfK+X7O4YZ1Vr14gfuhEpWImGUVEKZ1MqIcf6qoCKGlAbz4Ftv6g9CAy+X2LQgMrYK7AsSPDFwQ2IXLkpNB8885jAXlajrYNFi+BuHilCJjM=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDo385EdO3Q9a9gM7vB54U7tO40OJjhIKuCD/4Lczktc)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Saturday 21 March 2026  00:21:47 +0000 (0:00:01.013)       0:00:23.882 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCzB2yNwrdxiybJKIFTUFKw9hgA3eTgnpnh7Bnn2LQx2kmV9Tb3SUpq4hu3HGINP8vFI02QBpLJY3J8calXDdlmDA7s+90KFx+ituDeWizWVTuBtnDfWKH+bXqnpK4fytWz+rraBQg0zENXnkq95/sFYQAlEfklsEN6QtaTqx9UBFXWyST4nveGW9WLbfvqaBiGw67W1alnRV39FUhctaQRVoxfGLmERo7vX7PAhq+olypDXKriYYrJUJYej+TBdEhBQKbjHxl+ltGYYsCAgQ3YNMicYsutcKJq1fn5y1xhkGh2D/3jMSj8zE3H/ZbwyEuwMvRce6WxPHyfRwwEEcw1z+ZWS2fr565/aK/d+oa+dPgmHdAkOHTLTTw5GHE32YXuhZ+oqUORH5NyTeTtl0jj+JFCUt1rEN+4ZyNQQtIX4k4bxce/K7hFeAmBE3oxWq8RaaR95g89U5m1CmJ22hK1sCZ6+4kHW4gmhwZU0tVuv+dGZHtzxNdHQhnjrFnkEfk=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBMw16gD3Pbm/obmA1y0WkDRYOE6nz7IznqcYxNJs1Ttz0TfWx4+l2TZZB+cWEZlqH1wXHeH4gIujSHjYb0z47x4=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIPChxonu3Hcdz1NXmezrajxuZTmJk01UE7s701krwJ8B)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Saturday 21 March 2026  00:21:48 +0000 (0:00:01.032)       0:00:24.915 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAILcvUHyiPwzNQOC7IOncTN/c7TkgoFEUaxcoF1QveIN9)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC+2o4j0grHvLW3HhVtn4eHEakT9YYI3Sb++Ci0amagE5qNEbs/ZxyR9lzC65lYaEb3k10xr1I7dsrCB4xGyg89vnvU9BhFTChCpOBN4XSsQI0VXJcoK1gwh1vVcAgzoYbDh9qj25CpXEKmPS79tu5aMhzf7ENIjYP84/DyI7o/3zRk3GIocva42Ojjj8/EmmhNU0q+tfq4+kHnI6xCSoDK9YINST/sakm5VZWgzvZIluw7dbpbOvCgfflUmbaoO2emen9a1T23P/2RQLaCnVoiOij5+TIM4yZCPtFmg8Pf+BgRPt7dkjngX62KzF5NbXSNQGD7jI7L2bBWaOdKx1SmxbjIbQ/+ZwOcaUtfC49XfHHRrCQv6WXziDCPOSe+11xrCaHKcJMEztUEjJAR1Q0PI/bShfT+8TsULIMm1PYdKL+97cN57ccRrW2BEh1UExjG1LpvT+F0ihjouDQtOJdQTvGJHfAIwk0ttixM13ny9r/YvhZcZ8+OHpx75ljTc20=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLaFmk9jZnhiVtKL0wRcGE4qVsljEsur5kTWwGDIJ1BGW9608838oko7fKNiNcyELZFzn+EJxiyJc06R/YAnm94=)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Saturday 21 March 2026  00:21:49 +0000 (0:00:01.043)       0:00:25.958 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC8TUaJpxO/zwNXdrLifgopr+lsK/T9T7LNe2Pa4nLlZD69HGd8aTTUK8isoyIT+NnnT1Hna9SMLxNv8ibVyw6P7m3g8I6jzlc+J5ex443F1eIuwKlHiLKGrCCX+zp7DFbbJhcXDfl8cglHXPgdeyGVbDEWqOmWRSalh7QWAdF4dhVKp2J+rmuqJEz3MxqjNfpelItlm8Cd9dAL22kjI8YeqXTWW/q/5bkEwvApD5PqoxOe8FpSPtQUW1xx9/pwWyIbsN6ARcX8nrmzAHbwpHvc0GR39x1ahcLnFm+7UE+5Mxwq8SvCTQe2oeLkEg7XRdj3nJ7cUmGFTGczVMHULU6ZuHiabMhbwVbBBnjm5mfkCcOVdMOaYEAcqUBB51w/6RQpB+I23Di40MibLigFEoQOaIEHZtggAivOaXouqPlNQ21MevgBw1jzwVNtR44U4uBS2uMSrCeopIuxVYptnlOFa3YdDeS/gfAacyD3/dkLymC0rC4QBn1PjV46J6bbQ20=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBJcLJ530VGI+fZqYZivjtRjY60XsOdACkcMGZCO6DM1QCtAMis3Vt7RkgrBST8pYU1j4GVKr/b830ZCxBx8OWD8=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIIdSvMWwXHWpnHDy//kDiEa/cHDwBq24RvQjo19fNzjB)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write static known_hosts entries] ************",
                                "Saturday 21 March 2026  00:21:50 +0000 (0:00:01.043)       0:00:27.002 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write extra known_hosts entries] *************",
                                "Saturday 21 March 2026  00:21:50 +0000 (0:00:00.172)       0:00:27.175 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Delete known_hosts entries] ******************",
                                "Saturday 21 March 2026  00:21:50 +0000 (0:00:00.065)       0:00:27.240 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Set file permissions] ************************",
                                "Saturday 21 March 2026  00:21:51 +0000 (0:00:00.056)       0:00:27.296 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=31  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:21:51 +0000 (0:00:00.512)       0:00:27.809 ******** ",
                                "=============================================================================== ",
                                "osism.commons.known_hosts : Run ssh-keyscan for all hosts with hostname --- 6.44s",
                                "osism.commons.known_hosts : Run ssh-keyscan for all hosts with ansible_host --- 5.32s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.22s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.07s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.05s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.04s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.04s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.04s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.04s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.04s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.04s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.03s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.03s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.02s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.02s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.01s",
                                "osism.commons.known_hosts : Set file permissions ------------------------ 0.51s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with hostname --- 0.19s",
                                "osism.commons.known_hosts : Write static known_hosts entries ------------ 0.17s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with ansible_host --- 0.17s",
                                "+ osism apply squid",
                                "\u001b[32m2026-03-21 00:22:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of squid.\u001b[0m",
                                "\u001b[32m2026-03-21 00:22:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d030f99f-1670-411b-a7f7-ce57e9141b9b (squid) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 00:22:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task d030f99f-1670-411b-a7f7-ce57e9141b9b (squid) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role squid] ********************************************************",
                                "",
                                "TASK [osism.services.squid : Include install tasks] ****************************",
                                "Saturday 21 March 2026  00:22:06 +0000 (0:00:00.187)       0:00:00.187 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/squid/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Install required packages] ************************",
                                "Saturday 21 March 2026  00:22:06 +0000 (0:00:00.079)       0:00:00.267 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Create required directories] **********************",
                                "Saturday 21 March 2026  00:22:08 +0000 (0:00:02.335)       0:00:02.603 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/squid/configuration)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/squid/configuration/conf.d)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/squid)\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Copy squid configuration files] *******************",
                                "Saturday 21 March 2026  00:22:09 +0000 (0:00:01.168)       0:00:03.771 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=osism.conf)\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Remove osism_allow_list.conf configuration file] ***",
                                "Saturday 21 March 2026  00:22:10 +0000 (0:00:01.027)       0:00:04.798 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Copy docker-compose.yml file] *********************",
                                "Saturday 21 March 2026  00:22:11 +0000 (0:00:00.341)       0:00:05.140 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Manage squid service] *****************************",
                                "Saturday 21 March 2026  00:22:12 +0000 (0:00:00.884)       0:00:06.024 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage squid service (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.squid : Restart squid service] *****************",
                                "Saturday 21 March 2026  00:22:43 +0000 (0:00:30.911)       0:00:36.935 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.squid : Wait for squid service to start] *******",
                                "Saturday 21 March 2026  00:22:58 +0000 (0:00:15.784)       0:00:52.719 ******** ",
                                "Pausing for 60 seconds",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.squid : Register that squid service was restarted] ***",
                                "Saturday 21 March 2026  00:23:58 +0000 (0:01:00.084)       0:01:52.804 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.squid : Wait for an healthy squid service] *****",
                                "Saturday 21 March 2026  00:23:59 +0000 (0:00:00.060)       0:01:52.864 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=11  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:23:59 +0000 (0:00:00.585)       0:01:53.450 ******** ",
                                "=============================================================================== ",
                                "osism.services.squid : Wait for squid service to start ----------------- 60.08s",
                                "osism.services.squid : Manage squid service ---------------------------- 30.91s",
                                "osism.services.squid : Restart squid service --------------------------- 15.78s",
                                "osism.services.squid : Install required packages ------------------------ 2.34s",
                                "osism.services.squid : Create required directories ---------------------- 1.17s",
                                "osism.services.squid : Copy squid configuration files ------------------- 1.03s",
                                "osism.services.squid : Copy docker-compose.yml file --------------------- 0.88s",
                                "osism.services.squid : Wait for an healthy squid service ---------------- 0.59s",
                                "osism.services.squid : Remove osism_allow_list.conf configuration file --- 0.34s",
                                "osism.services.squid : Include install tasks ---------------------------- 0.08s",
                                "osism.services.squid : Register that squid service was restarted -------- 0.06s",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ /opt/configuration/scripts/set-kolla-namespace.sh kolla",
                                "+ set -e",
                                "+ NAMESPACE=kolla",
                                "+ sed -i 's#docker_namespace: .*#docker_namespace: kolla#g' /opt/configuration/inventory/group_vars/all/kolla.yml",
                                "++ semver latest 9.0.0",
                                "+ [[ -1 -lt 0 ]]",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ osism apply operator -u ubuntu -l testbed-nodes",
                                "\u001b[32m2026-03-21 00:24:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of operator.\u001b[0m",
                                "\u001b[32m2026-03-21 00:24:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ec44afcf-bc56-470f-92a0-18c80be44f09 (operator) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 00:24:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task ec44afcf-bc56-470f-92a0-18c80be44f09 (operator) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Make ssh pipelining working] *********************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Saturday 21 March 2026  00:24:14 +0000 (0:00:00.169)       0:00:00.169 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Do not require tty for all users] ****************************************",
                                "Saturday 21 March 2026  00:24:17 +0000 (0:00:03.447)       0:00:03.617 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role operator] *****************************************************",
                                "",
                                "TASK [osism.commons.operator : Gather variables for each operating system] *****",
                                "Saturday 21 March 2026  00:24:18 +0000 (0:00:00.916)       0:00:04.533 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set operator_groups variable to default value] ***",
                                "Saturday 21 March 2026  00:24:18 +0000 (0:00:00.157)       0:00:04.691 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Create operator group] **************************",
                                "Saturday 21 March 2026  00:24:19 +0000 (0:00:00.137)       0:00:04.828 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Create user] ************************************",
                                "Saturday 21 March 2026  00:24:19 +0000 (0:00:00.702)       0:00:05.530 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Add user to additional groups] ******************",
                                "Saturday 21 March 2026  00:24:20 +0000 (0:00:00.995)       0:00:06.526 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=sudo)\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Copy user sudoers file] *************************",
                                "Saturday 21 March 2026  00:24:22 +0000 (0:00:01.278)       0:00:07.805 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set language variables in .bashrc configuration file] ***",
                                "Saturday 21 March 2026  00:24:23 +0000 (0:00:01.335)       0:00:09.140 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module remote_tmp /root/.ansible/tmp did not exist and was created\u001b[0m",
                                "\u001b[1;35mwith a mode of 0700, this may cause issues when running as another user. To\u001b[0m",
                                "\u001b[1;35mavoid this, create the remote_tmp dir with the correct permissions manually\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set custom environment variables in .bashrc configuration file] ***",
                                "Saturday 21 March 2026  00:24:24 +0000 (0:00:01.251)       0:00:10.392 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file] ***",
                                "Saturday 21 March 2026  00:24:24 +0000 (0:00:00.144)       0:00:10.537 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Create .ssh directory] **************************",
                                "Saturday 21 March 2026  00:24:24 +0000 (0:00:00.151)       0:00:10.688 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Check number of SSH authorized keys] ************",
                                "Saturday 21 March 2026  00:24:25 +0000 (0:00:00.529)       0:00:11.218 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set ssh authorized keys] ************************",
                                "Saturday 21 March 2026  00:24:25 +0000 (0:00:00.183)       0:00:11.401 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Delete ssh authorized keys] *********************",
                                "Saturday 21 March 2026  00:24:26 +0000 (0:00:00.716)       0:00:12.118 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set authorized GitHub accounts] *****************",
                                "Saturday 21 March 2026  00:24:26 +0000 (0:00:00.144)       0:00:12.263 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Delete authorized GitHub accounts] **************",
                                "Saturday 21 March 2026  00:24:26 +0000 (0:00:00.144)       0:00:12.407 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set password] ***********************************",
                                "Saturday 21 March 2026  00:24:26 +0000 (0:00:00.133)       0:00:12.541 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Unset & lock password] **************************",
                                "Saturday 21 March 2026  00:24:27 +0000 (0:00:00.653)       0:00:13.194 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:24:27 +0000 (0:00:00.235)       0:00:13.429 ******** ",
                                "=============================================================================== ",
                                "Gathering Facts --------------------------------------------------------- 3.45s",
                                "osism.commons.operator : Copy user sudoers file ------------------------- 1.34s",
                                "osism.commons.operator : Add user to additional groups ------------------ 1.28s",
                                "osism.commons.operator : Set language variables in .bashrc configuration file --- 1.25s",
                                "osism.commons.operator : Create user ------------------------------------ 1.00s",
                                "Do not require tty for all users ---------------------------------------- 0.92s",
                                "osism.commons.operator : Set ssh authorized keys ------------------------ 0.72s",
                                "osism.commons.operator : Create operator group -------------------------- 0.70s",
                                "osism.commons.operator : Set password ----------------------------------- 0.65s",
                                "osism.commons.operator : Create .ssh directory -------------------------- 0.53s",
                                "osism.commons.operator : Unset & lock password -------------------------- 0.24s",
                                "osism.commons.operator : Check number of SSH authorized keys ------------ 0.18s",
                                "osism.commons.operator : Gather variables for each operating system ----- 0.16s",
                                "osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file --- 0.15s",
                                "osism.commons.operator : Delete ssh authorized keys --------------------- 0.14s",
                                "osism.commons.operator : Set custom environment variables in .bashrc configuration file --- 0.14s",
                                "osism.commons.operator : Set authorized GitHub accounts ----------------- 0.14s",
                                "osism.commons.operator : Set operator_groups variable to default value --- 0.14s",
                                "osism.commons.operator : Delete authorized GitHub accounts -------------- 0.13s",
                                "+ osism apply --environment custom facts",
                                "\u001b[32m2026-03-21 00:24:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play facts in environment custom\u001b[0m",
                                "\u001b[32m2026-03-21 00:24:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of facts.\u001b[0m",
                                "\u001b[32m2026-03-21 00:24:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9c35476d-6400-4b54-acfb-b274a3f36f36 (facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 00:24:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 9c35476d-6400-4b54-acfb-b274a3f36f36 (facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Copy custom network devices fact] ****************************************",
                                "",
                                "TASK [Create custom facts directory] *******************************************",
                                "Saturday 21 March 2026  00:24:42 +0000 (0:00:00.118)       0:00:00.118 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Copy fact file] **********************************************************",
                                "Saturday 21 March 2026  00:24:43 +0000 (0:00:01.350)       0:00:01.469 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY [Copy custom ceph devices facts] ******************************************",
                                "",
                                "TASK [osism.commons.repository : Gather variables for each operating system] ***",
                                "Saturday 21 March 2026  00:24:45 +0000 (0:00:01.345)       0:00:02.815 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Set repository_default fact to default value] ***",
                                "Saturday 21 March 2026  00:24:45 +0000 (0:00:00.096)       0:00:02.911 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Set repositories to default] ******************",
                                "Saturday 21 March 2026  00:24:45 +0000 (0:00:00.195)       0:00:03.106 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Include distribution specific repository tasks] ***",
                                "Saturday 21 March 2026  00:24:45 +0000 (0:00:00.194)       0:00:03.301 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****",
                                "Saturday 21 March 2026  00:24:45 +0000 (0:00:00.117)       0:00:03.419 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************",
                                "Saturday 21 March 2026  00:24:46 +0000 (0:00:00.425)       0:00:03.844 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Copy 99osism apt configuration] ***************",
                                "Saturday 21 March 2026  00:24:46 +0000 (0:00:00.115)       0:00:03.960 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Remove sources.list file] *********************",
                                "Saturday 21 March 2026  00:24:47 +0000 (0:00:01.062)       0:00:05.022 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Copy ubuntu.sources file] *********************",
                                "Saturday 21 March 2026  00:24:48 +0000 (0:00:00.496)       0:00:05.519 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Update package cache] *************************",
                                "Saturday 21 March 2026  00:24:49 +0000 (0:00:01.139)       0:00:06.659 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Install required packages (RedHat)] **************************************",
                                "Saturday 21 March 2026  00:25:06 +0000 (0:00:17.434)       0:00:24.093 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Install required packages (Debian)] **************************************",
                                "Saturday 21 March 2026  00:25:06 +0000 (0:00:00.091)       0:00:24.184 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create custom facts directory] *******************************************",
                                "Saturday 21 March 2026  00:25:13 +0000 (0:00:07.285)       0:00:31.469 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Copy fact files] *********************************************************",
                                "Saturday 21 March 2026  00:25:14 +0000 (0:00:00.487)       0:00:31.957 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_osd_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_osd_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_osd_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_osd_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_osd_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_osd_devices_all)\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.commons.repository : Force update of package cache] *****",
                                "Saturday 21 March 2026  00:25:18 +0000 (0:00:03.802)       0:00:35.759 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Saturday 21 March 2026  00:25:19 +0000 (0:00:01.551)       0:00:37.311 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:25:23 +0000 (0:00:03.837)       0:00:41.148 ******** ",
                                "=============================================================================== ",
                                "osism.commons.repository : Update package cache ------------------------ 17.43s",
                                "Install required packages (Debian) -------------------------------------- 7.29s",
                                "Gathers facts about hosts ----------------------------------------------- 3.84s",
                                "Copy fact files --------------------------------------------------------- 3.80s",
                                "osism.commons.repository : Force update of package cache ---------------- 1.55s",
                                "Create custom facts directory ------------------------------------------- 1.35s",
                                "Copy fact file ---------------------------------------------------------- 1.35s",
                                "osism.commons.repository : Copy ubuntu.sources file --------------------- 1.14s",
                                "osism.commons.repository : Copy 99osism apt configuration --------------- 1.06s",
                                "osism.commons.repository : Remove sources.list file --------------------- 0.50s",
                                "Create custom facts directory ------------------------------------------- 0.49s",
                                "osism.commons.repository : Create /etc/apt/sources.list.d directory ----- 0.43s",
                                "osism.commons.repository : Set repository_default fact to default value --- 0.20s",
                                "osism.commons.repository : Set repositories to default ------------------ 0.19s",
                                "osism.commons.repository : Include distribution specific repository tasks --- 0.12s",
                                "osism.commons.repository : Include tasks for Ubuntu < 24.04 ------------- 0.11s",
                                "osism.commons.repository : Gather variables for each operating system --- 0.10s",
                                "Install required packages (RedHat) -------------------------------------- 0.09s",
                                "+ osism apply bootstrap",
                                "\u001b[32m2026-03-21 00:25:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of bootstrap.\u001b[0m",
                                "\u001b[32m2026-03-21 00:25:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask adc203e2-da6f-4ccb-9c93-88f3a7415320 (bootstrap) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 00:25:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task adc203e2-da6f-4ccb-9c93-88f3a7415320 (bootstrap) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on state bootstrap] ************************************",
                                "",
                                "TASK [Group hosts based on state bootstrap] ************************************",
                                "Saturday 21 March 2026  00:25:38 +0000 (0:00:00.200)       0:00:00.200 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Saturday 21 March 2026  00:25:38 +0000 (0:00:00.303)       0:00:00.503 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts (if using --limit)] ***************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Saturday 21 March 2026  00:25:43 +0000 (0:00:04.886)       0:00:05.390 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Apply bootstrap roles part 1] ********************************************",
                                "",
                                "TASK [osism.commons.hostname : Set hostname] ***********************************",
                                "Saturday 21 March 2026  00:25:44 +0000 (0:00:00.449)       0:00:05.839 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.hostname : Copy /etc/hostname] *****************************",
                                "Saturday 21 March 2026  00:25:45 +0000 (0:00:01.460)       0:00:07.300 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.hosts : Include type specific tasks] ***********************",
                                "Saturday 21 March 2026  00:25:46 +0000 (0:00:01.249)       0:00:08.549 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/hosts/tasks/type-template.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.hosts : Copy /etc/hosts file] ******************************",
                                "Saturday 21 March 2026  00:25:47 +0000 (0:00:00.267)       0:00:08.817 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.proxy : Include distribution specific tasks] ***************",
                                "Saturday 21 March 2026  00:25:48 +0000 (0:00:01.489)       0:00:10.306 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/proxy/tasks/Debian-family.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.proxy : Configure proxy parameters for apt] ****************",
                                "Saturday 21 March 2026  00:25:49 +0000 (0:00:00.297)       0:00:10.603 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.proxy : Set system wide settings in environment file] ******",
                                "Saturday 21 March 2026  00:25:50 +0000 (0:00:01.030)       0:00:11.634 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.proxy : Remove system wide settings in environment file] ***",
                                "Saturday 21 March 2026  00:25:50 +0000 (0:00:00.593)       0:00:12.227 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Check minimum and maximum number of name servers] ***",
                                "Saturday 21 March 2026  00:25:51 +0000 (0:00:00.407)       0:00:12.635 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include resolvconf tasks] *********************",
                                "Saturday 21 March 2026  00:25:51 +0000 (0:00:00.186)       0:00:12.822 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-resolv.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include distribution specific installation tasks] ***",
                                "Saturday 21 March 2026  00:25:51 +0000 (0:00:00.284)       0:00:13.106 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf] ***",
                                "Saturday 21 March 2026  00:25:51 +0000 (0:00:00.293)       0:00:13.400 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Install package systemd-resolved] *************",
                                "Saturday 21 March 2026  00:25:53 +0000 (0:00:01.308)       0:00:14.708 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf] *****",
                                "Saturday 21 March 2026  00:25:53 +0000 (0:00:00.205)       0:00:14.914 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Archive existing file /etc/resolv.conf] *******",
                                "Saturday 21 March 2026  00:25:53 +0000 (0:00:00.589)       0:00:15.503 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf] ***",
                                "Saturday 21 March 2026  00:25:54 +0000 (0:00:00.227)       0:00:15.730 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Copy configuration files] *********************",
                                "Saturday 21 March 2026  00:25:54 +0000 (0:00:00.679)       0:00:16.410 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Start/enable systemd-resolved service] ********",
                                "Saturday 21 March 2026  00:25:56 +0000 (0:00:01.185)       0:00:17.595 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include distribution specific configuration tasks] ***",
                                "Saturday 21 March 2026  00:25:57 +0000 (0:00:01.017)       0:00:18.613 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Restart systemd-resolved service] *************",
                                "Saturday 21 March 2026  00:25:57 +0000 (0:00:00.330)       0:00:18.944 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Gather variables for each operating system] ***",
                                "Saturday 21 March 2026  00:25:58 +0000 (0:00:01.584)       0:00:20.529 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Set repository_default fact to default value] ***",
                                "Saturday 21 March 2026  00:25:59 +0000 (0:00:00.242)       0:00:20.771 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Set repositories to default] ******************",
                                "Saturday 21 March 2026  00:25:59 +0000 (0:00:00.252)       0:00:21.024 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Include distribution specific repository tasks] ***",
                                "Saturday 21 March 2026  00:25:59 +0000 (0:00:00.227)       0:00:21.252 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****",
                                "Saturday 21 March 2026  00:25:59 +0000 (0:00:00.320)       0:00:21.572 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************",
                                "Saturday 21 March 2026  00:26:00 +0000 (0:00:00.620)       0:00:22.193 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Copy 99osism apt configuration] ***************",
                                "Saturday 21 March 2026  00:26:00 +0000 (0:00:00.216)       0:00:22.410 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Remove sources.list file] *********************",
                                "Saturday 21 March 2026  00:26:01 +0000 (0:00:01.095)       0:00:23.506 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Copy ubuntu.sources file] *********************",
                                "Saturday 21 March 2026  00:26:02 +0000 (0:00:00.620)       0:00:24.126 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Update package cache] *************************",
                                "Saturday 21 March 2026  00:26:03 +0000 (0:00:01.204)       0:00:25.331 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Gather variables for each operating system] *****",
                                "Saturday 21 March 2026  00:26:20 +0000 (0:00:16.543)       0:00:41.874 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Set rsyslog_user variable to default value] *****",
                                "Saturday 21 March 2026  00:26:20 +0000 (0:00:00.250)       0:00:42.124 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Set rsyslog_workdir variable to default value] ***",
                                "Saturday 21 March 2026  00:26:20 +0000 (0:00:00.249)       0:00:42.374 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Include distribution specific install tasks] ****",
                                "Saturday 21 March 2026  00:26:21 +0000 (0:00:00.246)       0:00:42.620 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Install rsyslog package] ************************",
                                "Saturday 21 March 2026  00:26:21 +0000 (0:00:00.280)       0:00:42.901 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Copy rsyslog.conf configuration file] ***********",
                                "Saturday 21 March 2026  00:26:23 +0000 (0:00:02.151)       0:00:45.053 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Manage rsyslog service] *************************",
                                "Saturday 21 March 2026  00:26:24 +0000 (0:00:01.186)       0:00:46.239 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Include fluentd tasks] **************************",
                                "Saturday 21 March 2026  00:26:25 +0000 (0:00:00.978)       0:00:47.217 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/fluentd.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Forward syslog message to local fluentd daemon] ***",
                                "Saturday 21 March 2026  00:26:25 +0000 (0:00:00.324)       0:00:47.542 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Include additional log server tasks] ************",
                                "Saturday 21 March 2026  00:26:27 +0000 (0:00:01.196)       0:00:48.738 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Include logrotate tasks] ************************",
                                "Saturday 21 March 2026  00:26:27 +0000 (0:00:00.265)       0:00:49.004 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/logrotate.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Ensure logrotate package is installed] **********",
                                "Saturday 21 March 2026  00:26:27 +0000 (0:00:00.350)       0:00:49.354 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Configure logrotate for rsyslog] ****************",
                                "Saturday 21 March 2026  00:26:29 +0000 (0:00:01.931)       0:00:51.286 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.systohc : Install util-linux-extra package] ****************",
                                "Saturday 21 March 2026  00:26:30 +0000 (0:00:01.223)       0:00:52.509 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.systohc : Sync hardware clock] *****************************",
                                "Saturday 21 March 2026  00:26:42 +0000 (0:00:11.798)       0:01:04.308 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.configfs : Start sys-kernel-config mount] ******************",
                                "Saturday 21 March 2026  00:26:43 +0000 (0:00:00.939)       0:01:05.247 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Gather variables for each operating system] *****",
                                "Saturday 21 March 2026  00:26:44 +0000 (0:00:01.030)       0:01:06.278 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Set required_packages_distribution variable to default value] ***",
                                "Saturday 21 March 2026  00:26:44 +0000 (0:00:00.218)       0:01:06.497 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Include distribution specific package tasks] ****",
                                "Saturday 21 March 2026  00:26:45 +0000 (0:00:00.205)       0:01:06.702 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/packages/tasks/package-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Install needrestart package] ********************",
                                "Saturday 21 March 2026  00:26:45 +0000 (0:00:00.259)       0:01:06.962 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Set needrestart mode] ***************************",
                                "Saturday 21 March 2026  00:26:47 +0000 (0:00:01.953)       0:01:08.915 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Set apt_cache_valid_time variable to default value] ***",
                                "Saturday 21 March 2026  00:26:48 +0000 (0:00:00.730)       0:01:09.646 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Update package cache] ***************************",
                                "Saturday 21 March 2026  00:26:48 +0000 (0:00:00.304)       0:01:09.950 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Download upgrade packages] **********************",
                                "Saturday 21 March 2026  00:26:49 +0000 (0:00:01.379)       0:01:11.329 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Upgrade packages] *******************************",
                                "Saturday 21 March 2026  00:26:51 +0000 (0:00:02.211)       0:01:13.541 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Download required packages] *********************",
                                "Saturday 21 March 2026  00:26:54 +0000 (0:00:02.865)       0:01:16.406 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Install required packages] **********************",
                                "Saturday 21 March 2026  00:27:32 +0000 (0:00:37.683)       0:01:54.090 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Remove useless packages from the cache] *********",
                                "Saturday 21 March 2026  00:28:43 +0000 (0:01:10.988)       0:03:05.079 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Remove dependencies that are no longer required] ***",
                                "Saturday 21 March 2026  00:28:45 +0000 (0:00:01.862)       0:03:06.942 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Include sysctl tasks] *****************************",
                                "Saturday 21 March 2026  00:28:57 +0000 (0:00:11.961)       0:03:18.903 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 => (item={'key': 'elasticsearch', 'value': [{'name': 'vm.max_map_count', 'value': 262144}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 => (item={'key': 'rabbitmq', 'value': [{'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}, {'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}, {'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}, {'name': 'net.core.wmem_max', 'value': 16777216}, {'name': 'net.core.rmem_max', 'value': 16777216}, {'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}, {'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}, {'name': 'net.core.somaxconn', 'value': 4096}, {'name': 'net.ipv4.tcp_syncookies', 'value': 0}, {'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 => (item={'key': 'generic', 'value': [{'name': 'vm.swappiness', 'value': 1}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 => (item={'key': 'compute', 'value': [{'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 => (item={'key': 'network', 'value': [{'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 => (item={'key': 'k3s_node', 'value': [{'name': 'fs.inotify.max_user_instances', 'value': 1024}]})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on elasticsearch] ***********",
                                "Saturday 21 March 2026  00:28:57 +0000 (0:00:00.385)       0:03:19.289 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on rabbitmq] ****************",
                                "Saturday 21 March 2026  00:28:58 +0000 (0:00:00.604)       0:03:19.893 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on generic] *****************",
                                "Saturday 21 March 2026  00:29:04 +0000 (0:00:06.033)       0:03:25.927 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on compute] *****************",
                                "Saturday 21 March 2026  00:29:05 +0000 (0:00:01.634)       0:03:27.561 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on network] *****************",
                                "Saturday 21 March 2026  00:29:06 +0000 (0:00:00.599)       0:03:28.161 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on k3s_node] ****************",
                                "Saturday 21 March 2026  00:29:07 +0000 (0:00:00.536)       0:03:28.698 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m",
                                "",
                                "TASK [osism.commons.limits : Include limits tasks] *****************************",
                                "Saturday 21 March 2026  00:29:07 +0000 (0:00:00.717)       0:03:29.415 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.services : Populate service facts] *************************",
                                "Saturday 21 March 2026  00:29:08 +0000 (0:00:00.299)       0:03:29.715 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.services : Check services] *********************************",
                                "Saturday 21 March 2026  00:29:13 +0000 (0:00:05.164)       0:03:34.879 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.services : Start/enable required services] *****************",
                                "Saturday 21 March 2026  00:29:13 +0000 (0:00:00.289)       0:03:35.168 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=cron)\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Include distribution specific configure tasks] ******",
                                "Saturday 21 March 2026  00:29:14 +0000 (0:00:01.138)       0:03:36.307 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/motd/tasks/configure-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Remove update-motd package] *************************",
                                "Saturday 21 March 2026  00:29:15 +0000 (0:00:00.407)       0:03:36.714 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Check if /etc/default/motd-news exists] *************",
                                "Saturday 21 March 2026  00:29:16 +0000 (0:00:01.380)       0:03:38.094 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Disable the dynamic motd-news service] **************",
                                "Saturday 21 March 2026  00:29:17 +0000 (0:00:00.661)       0:03:38.756 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Get all configuration files in /etc/pam.d] **********",
                                "Saturday 21 March 2026  00:29:17 +0000 (0:00:00.676)       0:03:39.432 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Remove pam_motd.so rule] ****************************",
                                "Saturday 21 March 2026  00:29:18 +0000 (0:00:00.636)       0:03:40.068 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774051521.6864767, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774051567.6236367, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774051572.2527146, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774051533.1850278, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774051532.6675546, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774051566.528907, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1774051573.9143372, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Copy motd file] *************************************",
                                "Saturday 21 March 2026  00:29:19 +0000 (0:00:01.018)       0:03:41.087 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Copy issue file] ************************************",
                                "Saturday 21 March 2026  00:29:20 +0000 (0:00:01.094)       0:03:42.181 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Copy issue.net file] ********************************",
                                "Saturday 21 March 2026  00:29:21 +0000 (0:00:01.206)       0:03:43.387 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Configure SSH to print the motd] ********************",
                                "Saturday 21 March 2026  00:29:23 +0000 (0:00:01.252)       0:03:44.640 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Configure SSH to not print the motd] ****************",
                                "Saturday 21 March 2026  00:29:23 +0000 (0:00:00.258)       0:03:44.898 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.rng : Include distribution specific install tasks] ********",
                                "Saturday 21 March 2026  00:29:24 +0000 (0:00:00.861)       0:03:45.760 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rng/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.rng : Install rng package] ********************************",
                                "Saturday 21 March 2026  00:29:24 +0000 (0:00:00.411)       0:03:46.171 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.rng : Remove haveged package] *****************************",
                                "Saturday 21 March 2026  00:29:33 +0000 (0:00:08.446)       0:03:54.618 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.rng : Manage rng service] *********************************",
                                "Saturday 21 March 2026  00:29:34 +0000 (0:00:01.525)       0:03:56.144 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Gather variables for each operating system] ******",
                                "Saturday 21 March 2026  00:29:35 +0000 (0:00:01.228)       0:03:57.372 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Set cleanup_packages_distribution variable to default value] ***",
                                "Saturday 21 March 2026  00:29:36 +0000 (0:00:00.294)       0:03:57.667 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Set cleanup_services_distribution variable to default value] ***",
                                "Saturday 21 March 2026  00:29:36 +0000 (0:00:00.290)       0:03:57.957 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Populate service facts] **************************",
                                "Saturday 21 March 2026  00:29:36 +0000 (0:00:00.301)       0:03:58.259 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Include distribution specific timer tasks] *******",
                                "Saturday 21 March 2026  00:29:41 +0000 (0:00:04.876)       0:04:03.135 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/timers-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Disable apt-daily timers] ************************",
                                "Saturday 21 March 2026  00:29:41 +0000 (0:00:00.398)       0:04:03.533 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Include service tasks] ***************************",
                                "Saturday 21 March 2026  00:29:42 +0000 (0:00:00.392)       0:04:03.926 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/services-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Cleanup services] ********************************",
                                "Saturday 21 March 2026  00:29:42 +0000 (0:00:00.570)       0:04:04.496 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Include packages tasks] **************************",
                                "Saturday 21 March 2026  00:29:43 +0000 (0:00:00.350)       0:04:04.847 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/packages-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Cleanup installed packages] **********************",
                                "Saturday 21 March 2026  00:29:43 +0000 (0:00:00.456)       0:04:05.303 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Remove cloudinit package] ************************",
                                "Saturday 21 March 2026  00:30:15 +0000 (0:00:31.716)       0:04:37.020 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Uninstall unattended-upgrades package] ***********",
                                "Saturday 21 March 2026  00:30:24 +0000 (0:00:08.981)       0:04:46.002 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Remove useless packages from the cache] **********",
                                "Saturday 21 March 2026  00:30:33 +0000 (0:00:08.684)       0:04:54.686 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Remove dependencies that are no longer required] ***",
                                "Saturday 21 March 2026  00:30:34 +0000 (0:00:01.817)       0:04:56.504 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Include cloudinit tasks] *************************",
                                "Saturday 21 March 2026  00:30:41 +0000 (0:00:06.582)       0:05:03.086 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/cloudinit.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Remove cloud-init configuration directory] *******",
                                "Saturday 21 March 2026  00:30:41 +0000 (0:00:00.385)       0:05:03.472 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.timezone : Install tzdata package] *************************",
                                "Saturday 21 March 2026  00:30:42 +0000 (0:00:00.740)       0:05:04.212 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.timezone : Set timezone to UTC] ****************************",
                                "Saturday 21 March 2026  00:30:44 +0000 (0:00:01.786)       0:05:05.999 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.timezone : Create /etc/adjtime file] ***********************",
                                "Saturday 21 March 2026  00:30:45 +0000 (0:00:00.737)       0:05:06.737 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.timezone : Ensure UTC in /etc/adjtime] *********************",
                                "Saturday 21 March 2026  00:30:45 +0000 (0:00:00.271)       0:05:07.008 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Gather variables for each operating system] ******",
                                "Saturday 21 March 2026  00:30:45 +0000 (0:00:00.382)       0:05:07.390 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Set docker_version variable to default value] ****",
                                "Saturday 21 March 2026  00:30:46 +0000 (0:00:00.427)       0:05:07.818 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Set docker_cli_version variable to default value] ***",
                                "Saturday 21 March 2026  00:30:46 +0000 (0:00:00.278)       0:05:08.097 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Print used docker version] ***********************",
                                "Saturday 21 March 2026  00:30:46 +0000 (0:00:00.295)       0:05:08.393 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Print used docker cli version] *******************",
                                "Saturday 21 March 2026  00:30:47 +0000 (0:00:00.256)       0:05:08.649 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include block storage tasks] *********************",
                                "Saturday 21 March 2026  00:30:47 +0000 (0:00:00.266)       0:05:08.916 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include zram storage tasks] **********************",
                                "Saturday 21 March 2026  00:30:47 +0000 (0:00:00.238)       0:05:09.154 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include docker install tasks] ********************",
                                "Saturday 21 March 2026  00:30:47 +0000 (0:00:00.230)       0:05:09.385 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/install-docker-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Remove old architecture-dependent repository] ****",
                                "Saturday 21 March 2026  00:30:48 +0000 (0:00:00.387)       0:05:09.772 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Gather package facts] ****************************",
                                "Saturday 21 March 2026  00:30:49 +0000 (0:00:00.932)       0:05:10.705 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Check whether packages are installed that should not be installed] ***",
                                "Saturday 21 March 2026  00:30:52 +0000 (0:00:03.169)       0:05:13.874 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install apt-transport-https package] *************",
                                "Saturday 21 March 2026  00:30:52 +0000 (0:00:00.581)       0:05:14.456 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Add repository gpg key] **************************",
                                "Saturday 21 March 2026  00:30:59 +0000 (0:00:07.032)       0:05:21.488 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Add repository] **********************************",
                                "Saturday 21 March 2026  00:31:01 +0000 (0:00:01.158)       0:05:22.646 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Update package cache] ****************************",
                                "Saturday 21 March 2026  00:31:09 +0000 (0:00:08.294)       0:05:30.940 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Pin docker package version] **********************",
                                "Saturday 21 March 2026  00:31:12 +0000 (0:00:03.609)       0:05:34.550 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Pin docker-cli package version] ******************",
                                "Saturday 21 March 2026  00:31:14 +0000 (0:00:01.339)       0:05:35.889 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Unlock containerd package] ***********************",
                                "Saturday 21 March 2026  00:31:15 +0000 (0:00:01.380)       0:05:37.269 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install containerd package] **********************",
                                "Saturday 21 March 2026  00:31:16 +0000 (0:00:00.579)       0:05:37.849 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Lock containerd package] *************************",
                                "Saturday 21 March 2026  00:31:26 +0000 (0:00:10.532)       0:05:48.383 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install docker-cli package] **********************",
                                "Saturday 21 March 2026  00:31:27 +0000 (0:00:01.157)       0:05:49.540 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install docker package] **************************",
                                "Saturday 21 March 2026  00:31:37 +0000 (0:00:09.764)       0:05:59.305 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Unblock installation of python docker packages] ***",
                                "Saturday 21 March 2026  00:31:49 +0000 (0:00:11.395)       0:06:10.701 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=python-docker)\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install python3 docker package] ******************",
                                "Saturday 21 March 2026  00:31:50 +0000 (0:00:01.232)       0:06:11.933 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install python3 docker package from Debian Sid] ***",
                                "Saturday 21 March 2026  00:31:51 +0000 (0:00:00.660)       0:06:12.594 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Remove python docker packages (install python bindings from pip)] ***",
                                "Saturday 21 March 2026  00:31:55 +0000 (0:00:04.047)       0:06:16.641 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Block installation of python docker packages (install python bindings from pip)] ***",
                                "Saturday 21 March 2026  00:31:55 +0000 (0:00:00.505)       0:06:17.146 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install python3-pip package (install python bindings from pip)] ***",
                                "Saturday 21 March 2026  00:31:56 +0000 (0:00:00.542)       0:06:17.689 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install docker packages (install python bindings from pip)] ***",
                                "Saturday 21 March 2026  00:31:56 +0000 (0:00:00.479)       0:06:18.168 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install packages required by docker login] *******",
                                "Saturday 21 March 2026  00:31:57 +0000 (0:00:00.624)       0:06:18.793 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Ensure that some packages are not installed] *****",
                                "Saturday 21 March 2026  00:31:57 +0000 (0:00:00.525)       0:06:19.319 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include config tasks] ****************************",
                                "Saturday 21 March 2026  00:31:59 +0000 (0:00:01.941)       0:06:21.260 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/config.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Create plugins directory] ************************",
                                "Saturday 21 March 2026  00:32:00 +0000 (0:00:00.827)       0:06:22.088 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Create systemd overlay directory] ****************",
                                "Saturday 21 March 2026  00:32:01 +0000 (0:00:01.086)       0:06:23.174 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Copy systemd overlay file] ***********************",
                                "Saturday 21 March 2026  00:32:02 +0000 (0:00:00.853)       0:06:24.028 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Reload systemd daemon if systemd overlay file is changed] ***",
                                "Saturday 21 March 2026  00:32:03 +0000 (0:00:01.345)       0:06:25.373 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Copy limits configuration file] ******************",
                                "Saturday 21 March 2026  00:32:05 +0000 (0:00:01.372)       0:06:26.745 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Copy daemon.json configuration file] *************",
                                "Saturday 21 March 2026  00:32:06 +0000 (0:00:01.352)       0:06:28.097 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include service tasks] ***************************",
                                "Saturday 21 March 2026  00:32:08 +0000 (0:00:01.643)       0:06:29.740 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/service.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Reload systemd daemon] ***************************",
                                "Saturday 21 March 2026  00:32:08 +0000 (0:00:00.850)       0:06:30.591 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Manage service] **********************************",
                                "Saturday 21 March 2026  00:32:10 +0000 (0:00:01.381)       0:06:31.973 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Manage docker socket service] ********************",
                                "Saturday 21 March 2026  00:32:11 +0000 (0:00:01.283)       0:06:33.256 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Manage containerd service] ***********************",
                                "Saturday 21 March 2026  00:32:12 +0000 (0:00:01.149)       0:06:34.406 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include bootstrap tasks] *************************",
                                "Saturday 21 March 2026  00:32:13 +0000 (0:00:01.100)       0:06:35.506 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/bootstrap.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Saturday 21 March 2026  00:32:14 +0000 (0:00:00.918)       0:06:36.425 ******** ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Saturday 21 March 2026  00:32:14 +0000 (0:00:00.046)       0:06:36.472 ******** ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Saturday 21 March 2026  00:32:15 +0000 (0:00:00.242)       0:06:36.714 ******** ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Saturday 21 March 2026  00:32:15 +0000 (0:00:00.041)       0:06:36.756 ******** ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Saturday 21 March 2026  00:32:15 +0000 (0:00:00.041)       0:06:36.797 ******** ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Saturday 21 March 2026  00:32:15 +0000 (0:00:00.051)       0:06:36.848 ******** ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Saturday 21 March 2026  00:32:15 +0000 (0:00:00.041)       0:06:36.890 ******** ",
                                "",
                                "RUNNING HANDLER [osism.commons.repository : Force update of package cache] *****",
                                "Saturday 21 March 2026  00:32:15 +0000 (0:00:00.042)       0:06:36.932 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.rsyslog : Restart rsyslog service] *************",
                                "Saturday 21 March 2026  00:32:16 +0000 (0:00:01.336)       0:06:38.269 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.rsyslog : Restart logrotate service] ***********",
                                "Saturday 21 March 2026  00:32:17 +0000 (0:00:01.312)       0:06:39.581 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.docker : Restart docker service] ***************",
                                "Saturday 21 March 2026  00:32:19 +0000 (0:00:01.191)       0:06:40.773 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.docker : Wait after docker service restart] ****",
                                "Saturday 21 March 2026  00:32:21 +0000 (0:00:02.508)       0:06:43.281 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Add user to docker group] ************************",
                                "Saturday 21 March 2026  00:32:21 +0000 (0:00:00.103)       0:06:43.384 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Log into private registry and force re-authorization] ***",
                                "Saturday 21 March 2026  00:32:22 +0000 (0:00:01.180)       0:06:44.565 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include facts tasks] *****************************",
                                "Saturday 21 March 2026  00:32:23 +0000 (0:00:00.502)       0:06:45.067 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/facts.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Create facts directory] **************************",
                                "Saturday 21 March 2026  00:32:24 +0000 (0:00:00.890)       0:06:45.958 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Copy docker fact files] **************************",
                                "Saturday 21 March 2026  00:32:25 +0000 (0:00:01.205)       0:06:47.163 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=docker_images)\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : This install type is not supported] *******",
                                "Saturday 21 March 2026  00:32:28 +0000 (0:00:02.614)       0:06:49.778 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Include distribution specific install tasks] ***",
                                "Saturday 21 March 2026  00:32:28 +0000 (0:00:00.485)       0:06:50.264 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/docker_compose/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Remove docker-compose apt preferences file] ***",
                                "Saturday 21 March 2026  00:32:29 +0000 (0:00:00.926)       0:06:51.191 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Get checksum of docker-compose file] ******",
                                "Saturday 21 March 2026  00:32:30 +0000 (0:00:00.868)       0:06:52.059 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Remove docker-compose binary] *************",
                                "Saturday 21 March 2026  00:32:31 +0000 (0:00:00.832)       0:06:52.892 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Uninstall docker-compose package] *********",
                                "Saturday 21 March 2026  00:32:31 +0000 (0:00:00.526)       0:06:53.419 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Copy docker-compose script] ***************",
                                "Saturday 21 March 2026  00:32:33 +0000 (0:00:01.738)       0:06:55.157 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Install docker-compose-plugin package] ****",
                                "Saturday 21 March 2026  00:32:34 +0000 (0:00:00.779)       0:06:55.937 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Copy osism.target systemd file] ***********",
                                "Saturday 21 March 2026  00:32:42 +0000 (0:00:07.891)       0:07:03.828 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Enable osism.target] **********************",
                                "Saturday 21 March 2026  00:32:43 +0000 (0:00:01.348)       0:07:05.177 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Copy docker-compose systemd unit file] ****",
                                "Saturday 21 March 2026  00:32:45 +0000 (0:00:01.760)       0:07:06.937 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Create custom facts directory] *********************",
                                "Saturday 21 March 2026  00:32:47 +0000 (0:00:01.887)       0:07:08.825 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Copy fact files] ***********************************",
                                "Saturday 21 March 2026  00:32:48 +0000 (0:00:00.848)       0:07:09.673 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Check minimum and maximum number of servers] *****",
                                "Saturday 21 March 2026  00:32:48 +0000 (0:00:00.763)       0:07:10.436 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Gather variables for each operating system] ******",
                                "Saturday 21 March 2026  00:32:49 +0000 (0:00:00.674)       0:07:11.111 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Set chrony_conf_file variable to default value] ***",
                                "Saturday 21 March 2026  00:32:50 +0000 (0:00:00.513)       0:07:11.624 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Set chrony_key_file variable to default value] ***",
                                "Saturday 21 March 2026  00:32:50 +0000 (0:00:00.518)       0:07:12.143 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Populate service facts] **************************",
                                "Saturday 21 March 2026  00:32:51 +0000 (0:00:00.500)       0:07:12.643 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Manage timesyncd service] ************************",
                                "Saturday 21 March 2026  00:32:56 +0000 (0:00:05.565)       0:07:18.208 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Include distribution specific install tasks] *****",
                                "Saturday 21 March 2026  00:32:57 +0000 (0:00:00.673)       0:07:18.881 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Install package] *********************************",
                                "Saturday 21 March 2026  00:32:58 +0000 (0:00:00.794)       0:07:19.676 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Manage chrony service] ***************************",
                                "Saturday 21 March 2026  00:33:00 +0000 (0:00:01.952)       0:07:21.629 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Check if configuration file exists] **************",
                                "Saturday 21 March 2026  00:33:01 +0000 (0:00:01.444)       0:07:23.073 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Copy configuration file] *************************",
                                "Saturday 21 March 2026  00:33:02 +0000 (0:00:00.910)       0:07:23.983 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "",
                                "TASK [osism.services.lldpd : Include distribution specific install tasks] ******",
                                "Saturday 21 March 2026  00:33:04 +0000 (0:00:01.678)       0:07:25.662 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/lldpd/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.lldpd : Install lldpd package] ****************************",
                                "Saturday 21 March 2026  00:33:05 +0000 (0:00:00.938)       0:07:26.601 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.services.lldpd : Manage lldpd service] *****************************",
                                "Saturday 21 March 2026  00:33:14 +0000 (0:00:09.686)       0:07:36.287 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.commons.docker_compose : Reload systemd daemon] *********",
                                "Saturday 21 March 2026  00:33:16 +0000 (0:00:01.741)       0:07:38.029 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.chrony : Restart chrony service] ***************",
                                "Saturday 21 March 2026  00:33:17 +0000 (0:00:01.557)       0:07:39.587 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Apply bootstrap role part 2] *********************************************",
                                "",
                                "TASK [Include hardening role] **************************************************",
                                "Saturday 21 March 2026  00:33:19 +0000 (0:00:01.254)       0:07:40.841 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Apply bootstrap roles part 3] ********************************************",
                                "",
                                "TASK [osism.services.journald : Copy configuration file] ***********************",
                                "Saturday 21 March 2026  00:33:19 +0000 (0:00:00.512)       0:07:41.354 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.journald : Manage journald service] ***********************",
                                "Saturday 21 March 2026  00:33:21 +0000 (0:00:01.387)       0:07:42.741 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Include auditd role] *****************************************************",
                                "Saturday 21 March 2026  00:33:22 +0000 (0:00:01.627)       0:07:44.369 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Include smartd role] *****************************************************",
                                "Saturday 21 March 2026  00:33:23 +0000 (0:00:00.459)       0:07:44.829 ******** ",
                                "\u001b[0;36mincluded: osism.services.smartd for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Include distribution specific install tasks] *****",
                                "Saturday 21 March 2026  00:33:24 +0000 (0:00:00.791)       0:07:45.621 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/smartd/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Install smartmontools package] *******************",
                                "Saturday 21 March 2026  00:33:24 +0000 (0:00:00.921)       0:07:46.542 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Create /var/log/smartd directory] ****************",
                                "Saturday 21 March 2026  00:33:33 +0000 (0:00:09.044)       0:07:55.587 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Copy smartmontools configuration file] ***********",
                                "Saturday 21 March 2026  00:33:34 +0000 (0:00:00.838)       0:07:56.426 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Manage smartd service] ***************************",
                                "Saturday 21 March 2026  00:33:36 +0000 (0:00:01.372)       0:07:57.798 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.journald : Restart journald service] ***********",
                                "Saturday 21 March 2026  00:33:38 +0000 (0:00:01.881)       0:07:59.679 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.smartd : Restart smartd service] ***************",
                                "Saturday 21 March 2026  00:33:39 +0000 (0:00:01.202)       0:08:00.882 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Set state bootstrap] *****************************************************",
                                "",
                                "TASK [Set osism.bootstrap.status fact] *****************************************",
                                "Saturday 21 March 2026  00:33:40 +0000 (0:00:01.100)       0:08:01.983 ******** ",
                                "\u001b[0;36mincluded: osism.commons.state for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.state : Create custom facts directory] *********************",
                                "Saturday 21 March 2026  00:33:41 +0000 (0:00:01.091)       0:08:03.075 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.state : Write state into file] *****************************",
                                "Saturday 21 March 2026  00:33:42 +0000 (0:00:00.832)       0:08:03.907 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set osism.bootstrap.timestamp fact] **************************************",
                                "Saturday 21 March 2026  00:33:43 +0000 (0:00:01.245)       0:08:05.153 ******** ",
                                "\u001b[0;36mincluded: osism.commons.state for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.state : Create custom facts directory] *********************",
                                "Saturday 21 March 2026  00:33:44 +0000 (0:00:00.822)       0:08:05.976 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.state : Write state into file] *****************************",
                                "Saturday 21 March 2026  00:33:45 +0000 (0:00:00.831)       0:08:06.807 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=168 \u001b[0m \u001b[0;33mchanged=40  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=42  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=69  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=37  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=69  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=69  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=65  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=37  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=65  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=37  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=65  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=37  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:33:46 +0000 (0:00:01.312)       0:08:08.120 ******** ",
                                "=============================================================================== ",
                                "osism.commons.packages : Install required packages --------------------- 70.99s",
                                "osism.commons.packages : Download required packages -------------------- 37.68s",
                                "osism.commons.cleanup : Cleanup installed packages --------------------- 31.72s",
                                "osism.commons.repository : Update package cache ------------------------ 16.54s",
                                "osism.commons.packages : Remove dependencies that are no longer required -- 11.96s",
                                "osism.commons.systohc : Install util-linux-extra package --------------- 11.80s",
                                "osism.services.docker : Install docker package ------------------------- 11.40s",
                                "osism.services.docker : Install containerd package --------------------- 10.53s",
                                "osism.services.docker : Install docker-cli package ---------------------- 9.77s",
                                "osism.services.lldpd : Install lldpd package ---------------------------- 9.69s",
                                "osism.services.smartd : Install smartmontools package ------------------- 9.04s",
                                "osism.commons.cleanup : Remove cloudinit package ------------------------ 8.98s",
                                "osism.commons.cleanup : Uninstall unattended-upgrades package ----------- 8.68s",
                                "osism.services.rng : Install rng package -------------------------------- 8.45s",
                                "osism.services.docker : Add repository ---------------------------------- 8.29s",
                                "osism.commons.docker_compose : Install docker-compose-plugin package ---- 7.89s",
                                "osism.services.docker : Install apt-transport-https package ------------- 7.03s",
                                "osism.commons.cleanup : Remove dependencies that are no longer required --- 6.58s",
                                "osism.commons.sysctl : Set sysctl parameters on rabbitmq ---------------- 6.03s",
                                "osism.services.chrony : Populate service facts -------------------------- 5.57s",
                                "+ osism apply fail2ban",
                                "\u001b[32m2026-03-21 00:33:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of fail2ban.\u001b[0m",
                                "\u001b[32m2026-03-21 00:33:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 647b7b4d-9406-4744-8f65-bebc9111ce84 (fail2ban) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 00:33:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 647b7b4d-9406-4744-8f65-bebc9111ce84 (fail2ban) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role fail2ban] *****************************************************",
                                "",
                                "TASK [osism.services.fail2ban : Include distribution specific install tasks] ***",
                                "Saturday 21 March 2026  00:34:02 +0000 (0:00:00.330)       0:00:00.330 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/fail2ban/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.fail2ban : Install fail2ban package] **********************",
                                "Saturday 21 March 2026  00:34:03 +0000 (0:00:01.113)       0:00:01.443 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.services.fail2ban : Copy configuration files] **********************",
                                "Saturday 21 March 2026  00:34:15 +0000 (0:00:11.756)       0:00:13.200 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.fail2ban : Manage fail2ban service] ***********************",
                                "Saturday 21 March 2026  00:34:16 +0000 (0:00:01.613)       0:00:14.814 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.fail2ban : Reload fail2ban configuration] *****************",
                                "Saturday 21 March 2026  00:34:18 +0000 (0:00:01.244)       0:00:16.058 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:34:19 +0000 (0:00:01.520)       0:00:17.578 ******** ",
                                "=============================================================================== ",
                                "osism.services.fail2ban : Install fail2ban package --------------------- 11.76s",
                                "osism.services.fail2ban : Copy configuration files ---------------------- 1.61s",
                                "osism.services.fail2ban : Reload fail2ban configuration ----------------- 1.52s",
                                "osism.services.fail2ban : Manage fail2ban service ----------------------- 1.24s",
                                "osism.services.fail2ban : Include distribution specific install tasks --- 1.11s",
                                "+ [[ -e /etc/redhat-release ]]",
                                "+ osism apply network",
                                "\u001b[32m2026-03-21 00:34:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of network.\u001b[0m",
                                "\u001b[32m2026-03-21 00:34:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7cf0a656-9b93-4292-bfc5-7a3d9f2a5b10 (network) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 00:34:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 7cf0a656-9b93-4292-bfc5-7a3d9f2a5b10 (network) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role network] ******************************************************",
                                "",
                                "TASK [osism.commons.network : Gather variables for each operating system] ******",
                                "Saturday 21 March 2026  00:34:34 +0000 (0:00:00.295)       0:00:00.295 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include type specific tasks] *********************",
                                "Saturday 21 March 2026  00:34:34 +0000 (0:00:00.535)       0:00:00.831 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/netplan-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Install required packages] ***********************",
                                "Saturday 21 March 2026  00:34:35 +0000 (0:00:01.024)       0:00:01.856 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove ifupdown package] *************************",
                                "Saturday 21 March 2026  00:34:38 +0000 (0:00:02.491)       0:00:04.347 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Create required directories] *********************",
                                "Saturday 21 March 2026  00:34:40 +0000 (0:00:01.599)       0:00:05.946 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/etc/netplan)\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Prepare netplan configuration template] **********",
                                "Saturday 21 March 2026  00:34:41 +0000 (0:00:01.062)       0:00:07.009 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Copy netplan configuration] **********************",
                                "Saturday 21 March 2026  00:34:44 +0000 (0:00:02.968)       0:00:09.978 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove netplan configuration template] ***********",
                                "Saturday 21 March 2026  00:34:45 +0000 (0:00:01.492)       0:00:11.470 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Check if path for interface file exists] *********",
                                "Saturday 21 March 2026  00:34:47 +0000 (0:00:01.657)       0:00:13.128 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Copy interfaces file] ****************************",
                                "Saturday 21 March 2026  00:34:48 +0000 (0:00:00.892)       0:00:14.020 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Install package networkd-dispatcher] *************",
                                "Saturday 21 March 2026  00:34:48 +0000 (0:00:00.636)       0:00:14.657 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Copy dispatcher scripts] *************************",
                                "Saturday 21 March 2026  00:34:50 +0000 (0:00:02.003)       0:00:16.660 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'src': '/opt/configuration/network/iptables.sh', 'dest': 'routable.d/iptables.sh'})\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Manage service networkd-dispatcher] **************",
                                "Saturday 21 March 2026  00:34:51 +0000 (0:00:00.829)       0:00:17.490 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include cleanup tasks] ***************************",
                                "Saturday 21 March 2026  00:34:53 +0000 (0:00:01.550)       0:00:19.040 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/cleanup-netplan.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.network : List existing configuration files] ***************",
                                "Saturday 21 March 2026  00:34:54 +0000 (0:00:01.213)       0:00:20.253 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Set network_configured_files fact] ***************",
                                "Saturday 21 March 2026  00:34:55 +0000 (0:00:01.122)       0:00:21.375 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove unused configuration files] ***************",
                                "Saturday 21 March 2026  00:34:56 +0000 (0:00:00.697)       0:00:22.073 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include dummy interfaces] ************************",
                                "Saturday 21 March 2026  00:34:57 +0000 (0:00:00.941)       0:00:23.015 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include vxlan interfaces] ************************",
                                "Saturday 21 March 2026  00:34:57 +0000 (0:00:00.698)       0:00:23.713 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/vxlan-interfaces.yml for testbed-node-1, testbed-node-0, testbed-manager, testbed-node-4, testbed-node-5, testbed-node-3, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Create systemd networkd netdev files] ************",
                                "Saturday 21 March 2026  00:35:01 +0000 (0:00:03.941)       0:00:27.655 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.5', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'addresses': ['192.168.112.5/20']}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.13', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': []}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.11', 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': []}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.10', 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': []}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.5', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'addresses': ['192.168.128.5/20']}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.12', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': []}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.14', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'addresses': []}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.13', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': ['192.168.128.13/20']}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.15', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'addresses': []}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.11', 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': ['192.168.128.11/20']}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.10', 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': ['192.168.128.10/20']}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.12', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': ['192.168.128.12/20']}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.14', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'addresses': ['192.168.128.14/20']}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.15', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'addresses': ['192.168.128.15/20']}})\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Create systemd networkd network files] ***********",
                                "Saturday 21 March 2026  00:35:07 +0000 (0:00:05.363)       0:00:33.018 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.5', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'addresses': ['192.168.112.5/20']}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.10', 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': []}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.11', 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': []}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.14', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'addresses': []}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.5', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'addresses': ['192.168.128.5/20']}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.12', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': []}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.15', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'addresses': []}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.10', 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': ['192.168.128.10/20']}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan0', 'value': {'vni': 42, 'mtu': 1350, 'local_ip': '192.168.16.13', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': []}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.11', 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': ['192.168.128.11/20']}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.14', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'addresses': ['192.168.128.14/20']}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.12', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': ['192.168.128.12/20']}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.15', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'addresses': ['192.168.128.15/20']}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan1', 'value': {'vni': 23, 'mtu': 1350, 'local_ip': '192.168.16.13', 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'addresses': ['192.168.128.13/20']}})\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include networkd cleanup tasks] ******************",
                                "Saturday 21 March 2026  00:35:12 +0000 (0:00:05.248)       0:00:38.267 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/cleanup-networkd.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.network : List existing configuration files] ***************",
                                "Saturday 21 March 2026  00:35:13 +0000 (0:00:01.037)       0:00:39.305 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove unused configuration files] ***************",
                                "Saturday 21 March 2026  00:35:14 +0000 (0:00:00.989)       0:00:40.294 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include network extra init] **********************",
                                "Saturday 21 March 2026  00:35:15 +0000 (0:00:00.645)       0:00:40.940 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/network-extra-init.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Deploy network-extra-init script] ****************",
                                "Saturday 21 March 2026  00:35:16 +0000 (0:00:01.054)       0:00:41.994 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Deploy network-extra-init systemd service] *******",
                                "Saturday 21 March 2026  00:35:16 +0000 (0:00:00.630)       0:00:42.625 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Enable and start network-extra-init service] *****",
                                "Saturday 21 March 2026  00:35:17 +0000 (0:00:00.521)       0:00:43.146 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Disable and stop network-extra-init service] *****",
                                "Saturday 21 March 2026  00:35:17 +0000 (0:00:00.642)       0:00:43.789 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove network-extra-init systemd service] *******",
                                "Saturday 21 March 2026  00:35:19 +0000 (0:00:01.469)       0:00:45.258 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove network-extra-init script] ****************",
                                "Saturday 21 March 2026  00:35:20 +0000 (0:00:01.043)       0:00:46.301 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.commons.network : Reload systemd-networkd] **************",
                                "Saturday 21 March 2026  00:35:22 +0000 (0:00:02.139)       0:00:48.441 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.commons.network : Netplan configuration changed] ********",
                                "Saturday 21 March 2026  00:35:23 +0000 (0:00:00.602)       0:00:49.043 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=25  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:35:23 +0000 (0:00:00.671)       0:00:49.715 ******** ",
                                "=============================================================================== ",
                                "osism.commons.network : Create systemd networkd netdev files ------------ 5.36s",
                                "osism.commons.network : Create systemd networkd network files ----------- 5.25s",
                                "osism.commons.network : Include vxlan interfaces ------------------------ 3.94s",
                                "osism.commons.network : Prepare netplan configuration template ---------- 2.97s",
                                "osism.commons.network : Install required packages ----------------------- 2.49s",
                                "osism.commons.network : Remove network-extra-init script ---------------- 2.14s",
                                "osism.commons.network : Install package networkd-dispatcher ------------- 2.00s",
                                "osism.commons.network : Remove netplan configuration template ----------- 1.66s",
                                "osism.commons.network : Remove ifupdown package ------------------------- 1.60s",
                                "osism.commons.network : Manage service networkd-dispatcher -------------- 1.55s",
                                "osism.commons.network : Copy netplan configuration ---------------------- 1.49s",
                                "osism.commons.network : Disable and stop network-extra-init service ----- 1.47s",
                                "osism.commons.network : Include cleanup tasks --------------------------- 1.21s",
                                "osism.commons.network : List existing configuration files --------------- 1.12s",
                                "osism.commons.network : Create required directories --------------------- 1.06s",
                                "osism.commons.network : Include network extra init ---------------------- 1.05s",
                                "osism.commons.network : Remove network-extra-init systemd service ------- 1.04s",
                                "osism.commons.network : Include networkd cleanup tasks ------------------ 1.04s",
                                "osism.commons.network : Include type specific tasks --------------------- 1.02s",
                                "osism.commons.network : List existing configuration files --------------- 0.99s",
                                "+ osism apply wireguard",
                                "\u001b[32m2026-03-21 00:35:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of wireguard.\u001b[0m",
                                "\u001b[32m2026-03-21 00:35:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 101272ea-a17c-41c2-a57e-76286cdec110 (wireguard) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 00:35:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 101272ea-a17c-41c2-a57e-76286cdec110 (wireguard) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role wireguard] ****************************************************",
                                "",
                                "TASK [osism.services.wireguard : Install iptables package] *********************",
                                "Saturday 21 March 2026  00:35:38 +0000 (0:00:00.330)       0:00:00.330 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Install wireguard package] ********************",
                                "Saturday 21 March 2026  00:35:40 +0000 (0:00:01.760)       0:00:02.090 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Create public and private key - server] *******",
                                "Saturday 21 March 2026  00:35:46 +0000 (0:00:06.105)       0:00:08.196 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Create preshared key] *************************",
                                "Saturday 21 March 2026  00:35:47 +0000 (0:00:00.522)       0:00:08.718 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Get preshared key] ****************************",
                                "Saturday 21 March 2026  00:35:47 +0000 (0:00:00.419)       0:00:09.138 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Get public key - server] **********************",
                                "Saturday 21 March 2026  00:35:48 +0000 (0:00:00.559)       0:00:09.698 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Get private key - server] *********************",
                                "Saturday 21 March 2026  00:35:48 +0000 (0:00:00.399)       0:00:10.097 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Copy wg0.conf configuration file] *************",
                                "Saturday 21 March 2026  00:35:49 +0000 (0:00:00.406)       0:00:10.505 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Copy client configuration files] **************",
                                "Saturday 21 March 2026  00:35:50 +0000 (0:00:01.157)       0:00:11.663 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Manage wg-quick@wg0.service service] **********",
                                "Saturday 21 March 2026  00:35:51 +0000 (0:00:00.914)       0:00:12.577 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.wireguard : Restart wg0 service] ***************",
                                "Saturday 21 March 2026  00:35:53 +0000 (0:00:01.886)       0:00:14.464 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=11  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:35:53 +0000 (0:00:00.868)       0:00:15.333 ******** ",
                                "=============================================================================== ",
                                "osism.services.wireguard : Install wireguard package -------------------- 6.11s",
                                "osism.services.wireguard : Manage wg-quick@wg0.service service ---------- 1.89s",
                                "osism.services.wireguard : Install iptables package --------------------- 1.76s",
                                "osism.services.wireguard : Copy wg0.conf configuration file ------------- 1.16s",
                                "osism.services.wireguard : Copy client configuration files -------------- 0.91s",
                                "osism.services.wireguard : Restart wg0 service -------------------------- 0.87s",
                                "osism.services.wireguard : Get preshared key ---------------------------- 0.56s",
                                "osism.services.wireguard : Create public and private key - server ------- 0.52s",
                                "osism.services.wireguard : Create preshared key ------------------------- 0.42s",
                                "osism.services.wireguard : Get private key - server --------------------- 0.41s",
                                "osism.services.wireguard : Get public key - server ---------------------- 0.40s",
                                "+ sh -c /opt/configuration/scripts/prepare-wireguard-configuration.sh",
                                "  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current",
                                "                                 Dload  Upload   Total   Spent    Left  Speed",
                                "",
                                "  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0",
                                "100    15  100    15    0     0    202      0 --:--:-- --:--:-- --:--:--   205",
                                "+ osism apply --environment custom workarounds",
                                "\u001b[32m2026-03-21 00:35:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play workarounds in environment custom\u001b[0m",
                                "\u001b[32m2026-03-21 00:36:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of workarounds.\u001b[0m",
                                "\u001b[32m2026-03-21 00:36:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 693fe854-71d3-49da-b5ea-134491721a6f (workarounds) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 00:36:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 693fe854-71d3-49da-b5ea-134491721a6f (workarounds) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on virtualization_role] ********************************",
                                "Saturday 21 March 2026  00:36:08 +0000 (0:00:00.175)       0:00:00.175 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=virtualization_role_guest)\u001b[0m",
                                "",
                                "PLAY [Apply netplan configuration on the manager node] *************************",
                                "",
                                "TASK [Apply netplan configuration] *********************************************",
                                "Saturday 21 March 2026  00:36:09 +0000 (0:00:00.778)       0:00:00.953 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Apply netplan configuration on all other nodes] **************************",
                                "",
                                "TASK [Apply netplan configuration] *********************************************",
                                "Saturday 21 March 2026  00:36:12 +0000 (0:00:02.584)       0:00:03.537 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Add custom CA certificates to non-manager nodes] *************************",
                                "",
                                "TASK [Copy custom CA certificates] *********************************************",
                                "Saturday 21 March 2026  00:36:14 +0000 (0:00:02.415)       0:00:05.953 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "",
                                "TASK [Run update-ca-certificates] **********************************************",
                                "Saturday 21 March 2026  00:36:15 +0000 (0:00:01.390)       0:00:07.343 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Run update-ca-trust] *****************************************************",
                                "Saturday 21 March 2026  00:36:19 +0000 (0:00:03.938)       0:00:11.282 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Add a workaround service] ************************************************",
                                "",
                                "TASK [Copy workarounds.sh scripts] *********************************************",
                                "Saturday 21 March 2026  00:36:20 +0000 (0:00:00.546)       0:00:11.828 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Copy workarounds systemd unit file] **************************************",
                                "Saturday 21 March 2026  00:36:22 +0000 (0:00:01.760)       0:00:13.589 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Reload systemd daemon] ***************************************************",
                                "Saturday 21 March 2026  00:36:23 +0000 (0:00:01.489)       0:00:15.079 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Enable workarounds.service (Debian)] *************************************",
                                "Saturday 21 March 2026  00:36:25 +0000 (0:00:01.604)       0:00:16.683 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Enable and start workarounds.service (RedHat)] ***************************",
                                "Saturday 21 March 2026  00:36:26 +0000 (0:00:01.567)       0:00:18.250 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [On Ubuntu 24.04 install python3-docker from Debian Sid] ******************",
                                "",
                                "TASK [Install python3-docker] **************************************************",
                                "Saturday 21 March 2026  00:36:27 +0000 (0:00:00.739)       0:00:18.990 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:36:30 +0000 (0:00:02.629)       0:00:21.619 ******** ",
                                "=============================================================================== ",
                                "Run update-ca-certificates ---------------------------------------------- 3.94s",
                                "Install python3-docker -------------------------------------------------- 2.63s",
                                "Apply netplan configuration --------------------------------------------- 2.58s",
                                "Apply netplan configuration --------------------------------------------- 2.42s",
                                "Copy workarounds.sh scripts --------------------------------------------- 1.76s",
                                "Reload systemd daemon --------------------------------------------------- 1.60s",
                                "Enable workarounds.service (Debian) ------------------------------------- 1.57s",
                                "Copy workarounds systemd unit file -------------------------------------- 1.49s",
                                "Copy custom CA certificates --------------------------------------------- 1.39s",
                                "Group hosts based on virtualization_role -------------------------------- 0.78s",
                                "Enable and start workarounds.service (RedHat) --------------------------- 0.74s",
                                "Run update-ca-trust ----------------------------------------------------- 0.55s",
                                "+ osism apply reboot -l testbed-nodes -e ireallymeanit=yes",
                                "\u001b[32m2026-03-21 00:36:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of reboot.\u001b[0m",
                                "\u001b[32m2026-03-21 00:36:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8dda0dcf-3289-4ca7-9fdd-cfa365aeff77 (reboot) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 00:36:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 8dda0dcf-3289-4ca7-9fdd-cfa365aeff77 (reboot) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Saturday 21 March 2026  00:36:44 +0000 (0:00:00.223)       0:00:00.223 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Saturday 21 March 2026  00:36:44 +0000 (0:00:00.126)       0:00:00.350 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Saturday 21 March 2026  00:36:46 +0000 (0:00:01.241)       0:00:01.591 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Saturday 21 March 2026  00:36:46 +0000 (0:00:00.099)       0:00:01.690 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Saturday 21 March 2026  00:36:46 +0000 (0:00:00.086)       0:00:01.777 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Saturday 21 March 2026  00:36:47 +0000 (0:00:01.009)       0:00:02.787 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Saturday 21 March 2026  00:36:47 +0000 (0:00:00.114)       0:00:02.901 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Saturday 21 March 2026  00:36:47 +0000 (0:00:00.094)       0:00:02.996 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Saturday 21 March 2026  00:36:48 +0000 (0:00:01.035)       0:00:04.031 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Saturday 21 March 2026  00:36:48 +0000 (0:00:00.109)       0:00:04.141 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Saturday 21 March 2026  00:36:48 +0000 (0:00:00.094)       0:00:04.235 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Saturday 21 March 2026  00:36:49 +0000 (0:00:01.024)       0:00:05.259 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Saturday 21 March 2026  00:36:49 +0000 (0:00:00.109)       0:00:05.369 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Saturday 21 March 2026  00:36:50 +0000 (0:00:00.203)       0:00:05.573 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Saturday 21 March 2026  00:36:51 +0000 (0:00:01.000)       0:00:06.573 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Saturday 21 March 2026  00:36:51 +0000 (0:00:00.111)       0:00:06.685 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Saturday 21 March 2026  00:36:51 +0000 (0:00:00.096)       0:00:06.781 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Saturday 21 March 2026  00:36:52 +0000 (0:00:00.999)       0:00:07.781 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:36:52 +0000 (0:00:00.033)       0:00:07.814 ******** ",
                                "=============================================================================== ",
                                "Reboot system - do not wait for the reboot to complete ------------------ 6.31s",
                                "Exit playbook, if user did not mean to reboot systems ------------------- 0.70s",
                                "Reboot system - wait for the reboot to complete ------------------------- 0.58s",
                                "+ osism apply wait-for-connection -l testbed-nodes -e ireallymeanit=yes",
                                "\u001b[32m2026-03-21 00:37:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of wait-for-connection.\u001b[0m",
                                "\u001b[32m2026-03-21 00:37:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d0bc427e-2de9-41e8-a8db-b0f3e60a1221 (wait-for-connection) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 00:37:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task d0bc427e-2de9-41e8-a8db-b0f3e60a1221 (wait-for-connection) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Wait until remote systems are reachable] *********************************",
                                "",
                                "TASK [Wait until remote system is reachable] ***********************************",
                                "Saturday 21 March 2026  00:37:07 +0000 (0:00:00.310)       0:00:00.310 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:37:19 +0000 (0:00:11.510)       0:00:11.821 ******** ",
                                "=============================================================================== ",
                                "Wait until remote system is reachable ---------------------------------- 11.51s",
                                "+ osism apply hddtemp",
                                "\u001b[32m2026-03-21 00:37:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of hddtemp.\u001b[0m",
                                "\u001b[32m2026-03-21 00:37:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f636c49-3ee5-46f3-b998-ad491cd36695 (hddtemp) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 00:37:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 3f636c49-3ee5-46f3-b998-ad491cd36695 (hddtemp) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role hddtemp] ******************************************************",
                                "",
                                "TASK [osism.services.hddtemp : Gather variables for each operating system] *****",
                                "Saturday 21 March 2026  00:37:33 +0000 (0:00:00.281)       0:00:00.281 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Include distribution specific install tasks] ****",
                                "Saturday 21 March 2026  00:37:34 +0000 (0:00:00.543)       0:00:00.825 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/hddtemp/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Remove hddtemp package] *************************",
                                "Saturday 21 March 2026  00:37:35 +0000 (0:00:01.039)       0:00:01.864 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Enable Kernel Module drivetemp] *****************",
                                "Saturday 21 March 2026  00:37:37 +0000 (0:00:02.460)       0:00:04.324 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Check if drivetemp module is available] *********",
                                "Saturday 21 March 2026  00:37:38 +0000 (0:00:01.047)       0:00:05.371 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Load Kernel Module drivetemp] *******************",
                                "Saturday 21 March 2026  00:37:40 +0000 (0:00:01.334)       0:00:06.705 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Install lm-sensors] *****************************",
                                "Saturday 21 March 2026  00:37:40 +0000 (0:00:00.607)       0:00:07.313 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Include distribution specific service tasks] ****",
                                "Saturday 21 March 2026  00:37:54 +0000 (0:00:13.421)       0:00:20.735 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/hddtemp/tasks/service-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Manage lm-sensors service] **********************",
                                "Saturday 21 March 2026  00:37:55 +0000 (0:00:01.050)       0:00:21.785 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:37:57 +0000 (0:00:02.460)       0:00:24.246 ******** ",
                                "=============================================================================== ",
                                "osism.services.hddtemp : Install lm-sensors ---------------------------- 13.42s",
                                "osism.services.hddtemp : Manage lm-sensors service ---------------------- 2.46s",
                                "osism.services.hddtemp : Remove hddtemp package ------------------------- 2.46s",
                                "osism.services.hddtemp : Check if drivetemp module is available --------- 1.33s",
                                "osism.services.hddtemp : Include distribution specific service tasks ---- 1.05s",
                                "osism.services.hddtemp : Enable Kernel Module drivetemp ----------------- 1.05s",
                                "osism.services.hddtemp : Include distribution specific install tasks ---- 1.04s",
                                "osism.services.hddtemp : Load Kernel Module drivetemp ------------------- 0.61s",
                                "osism.services.hddtemp : Gather variables for each operating system ----- 0.54s",
                                "++ semver latest 7.1.1",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ sudo systemctl restart manager.service",
                                "+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]",
                                "+ wait_for_container_healthy 60 ceph-ansible",
                                "+ local max_attempts=60",
                                "+ local name=ceph-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ wait_for_container_healthy 60 kolla-ansible",
                                "+ local max_attempts=60",
                                "+ local name=kolla-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' kolla-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ wait_for_container_healthy 60 osism-ansible",
                                "+ local max_attempts=60",
                                "+ local name=osism-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' osism-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ [[ true == \\t\\r\\u\\e ]]",
                                "+ sh -c /opt/configuration/scripts/disable-ara.sh",
                                "ARA in ceph-ansible already disabled.",
                                "ARA in kolla-ansible already disabled.",
                                "ARA in osism-ansible already disabled.",
                                "ARA in osism-kubernetes already disabled.",
                                "+ osism apply gather-facts",
                                "\u001b[32m2026-03-21 00:39:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of gather-facts.\u001b[0m",
                                "\u001b[32m2026-03-21 00:39:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 25cfd3ec-85e1-4c1c-9913-f49fc78dae3a (gather-facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 00:39:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 25cfd3ec-85e1-4c1c-9913-f49fc78dae3a (gather-facts) has been started and output is visible here.\u001b[0m",
                                "\u001b[1;35m[WARNING]: Invalid characters were found in group names but not replaced, use\u001b[0m",
                                "\u001b[1;35m-vvvv to see details\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "\u001b[1;31mfatal: [testbed-manager]: UNREACHABLE! => {\"changed\": false, \"msg\": \"Data could not be sent to remote host \\\"192.168.16.5\\\". Make sure this host can be reached over ssh: no such identity: /ansible/secrets/id_rsa: No such file or directory\\r\\ndragon@192.168.16.5: Permission denied (publickey).\\r\\n\", \"unreachable\": true}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "\u001b[1;31mfatal: [testbed-node-5]: UNREACHABLE! => {\"changed\": false, \"msg\": \"Data could not be sent to remote host \\\"192.168.16.15\\\". Make sure this host can be reached over ssh: no such identity: /ansible/secrets/id_rsa: No such file or directory\\r\\ndragon@192.168.16.15: Permission denied (publickey).\\r\\n\", \"unreachable\": true}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "\u001b[1;31mfatal: [testbed-node-1]: UNREACHABLE! => {\"changed\": false, \"msg\": \"Data could not be sent to remote host \\\"192.168.16.11\\\". Make sure this host can be reached over ssh: no such identity: /ansible/secrets/id_rsa: No such file or directory\\r\\ndragon@192.168.16.11: Permission denied (publickey).\\r\\n\", \"unreachable\": true}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "\u001b[1;31mfatal: [testbed-node-2]: UNREACHABLE! => {\"changed\": false, \"msg\": \"Data could not be sent to remote host \\\"192.168.16.12\\\". Make sure this host can be reached over ssh: no such identity: /ansible/secrets/id_rsa: No such file or directory\\r\\ndragon@192.168.16.12: Permission denied (publickey).\\r\\n\", \"unreachable\": true}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "\u001b[1;31mfatal: [testbed-node-3]: UNREACHABLE! => {\"changed\": false, \"msg\": \"Data could not be sent to remote host \\\"192.168.16.13\\\". Make sure this host can be reached over ssh: no such identity: /ansible/secrets/id_rsa: No such file or directory\\r\\ndragon@192.168.16.13: Permission denied (publickey).\\r\\n\", \"unreachable\": true}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "\u001b[1;31mfatal: [testbed-node-0]: UNREACHABLE! => {\"changed\": false, \"msg\": \"Data could not be sent to remote host \\\"192.168.16.10\\\". Make sure this host can be reached over ssh: no such identity: /ansible/secrets/id_rsa: No such file or directory\\r\\ndragon@192.168.16.10: Permission denied (publickey).\\r\\n\", \"unreachable\": true}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "\u001b[1;31mfatal: [testbed-node-4]: UNREACHABLE! => {\"changed\": false, \"msg\": \"Data could not be sent to remote host \\\"192.168.16.14\\\". Make sure this host can be reached over ssh: no such identity: /ansible/secrets/id_rsa: No such file or directory\\r\\ndragon@192.168.16.14: Permission denied (publickey).\\r\\n\", \"unreachable\": true}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts if using --limit] *****************************",
                                "",
                                "TASK [Gather facts for all hosts] **********************************************",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/001-helpers.sh /usr/local/bin/deploy-helper",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/100-ceph-with-ansible.sh /usr/local/bin/deploy-ceph-with-ansible",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/100-ceph-with-rook.sh /usr/local/bin/deploy-ceph-with-rook",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/200-infrastructure.sh /usr/local/bin/deploy-infrastructure",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/300-openstack.sh /usr/local/bin/deploy-openstack",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/320-openstack-minimal.sh /usr/local/bin/deploy-openstack-minimal",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/400-monitoring.sh /usr/local/bin/deploy-monitoring",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/500-kubernetes.sh /usr/local/bin/deploy-kubernetes",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/510-clusterapi.sh /usr/local/bin/deploy-kubernetes-clusterapi",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade-manager.sh /usr/local/bin/upgrade-manager",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/100-ceph-with-ansible.sh /usr/local/bin/upgrade-ceph-with-ansible",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/100-ceph-with-rook.sh /usr/local/bin/upgrade-ceph-with-rook",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/200-infrastructure.sh /usr/local/bin/upgrade-infrastructure",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/300-openstack.sh /usr/local/bin/upgrade-openstack",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/320-openstack-minimal.sh /usr/local/bin/upgrade-openstack-minimal",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/400-monitoring.sh /usr/local/bin/upgrade-monitoring",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/500-kubernetes.sh /usr/local/bin/upgrade-kubernetes",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/510-clusterapi.sh /usr/local/bin/upgrade-kubernetes-clusterapi",
                                "+ sudo ln -sf /opt/configuration/scripts/bootstrap/300-openstack.sh /usr/local/bin/bootstrap-openstack",
                                "+ sudo ln -sf /opt/configuration/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh /usr/local/bin/bootstrap-octavia",
                                "+ sudo ln -sf /opt/configuration/scripts/bootstrap/302-openstack-k8s-clusterapi-images.sh /usr/local/bin/bootstrap-clusterapi",
                                "+ sudo ln -sf /opt/configuration/scripts/disable-local-registry.sh /usr/local/bin/disable-local-registry",
                                "+ sudo ln -sf /opt/configuration/scripts/pull-images.sh /usr/local/bin/pull-images",
                                "+ [[ false == \\t\\r\\u\\e ]]"
                            ],
                            "zuul_log_id": "0242ac1f-6606-9900-9c98-00000000001a-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:39:33.446917Z",
                            "start": "2026-03-21T00:15:56.875499Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-00000000001a",
                        "name": "Deploy manager + bootstrap nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": false,
                            "false_condition": "not _nutshell | bool",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T00:39:34.049033Z",
                            "start": "2026-03-21T00:39:33.540088Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-00000000001b",
                        "name": "Deploy services"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed04/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.193.136",
                                "/opt/configuration/scripts/deploy-in-a-nutshell.sh"
                            ],
                            "delta": "0:32:42.091309",
                            "end": "2026-03-21 01:12:16.461951",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed04/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.193.136 /opt/configuration/scripts/deploy-in-a-nutshell.sh",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-9900-9c98-00000000001c-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-21 00:39:34.370642",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n\n# PULL IMAGES\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=true\n++ TEMPEST=true\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ echo\n+ echo '# PULL IMAGES'\n\n+ echo\n++ semver latest 7.0.0\n+ [[ -1 -ge 0 ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ osism apply --no-wait -r 2 -e custom pull-images\n\u001b[32m2026-03-21 00:39:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play pull-images in environment custom\u001b[0m\n\u001b[32m2026-03-21 00:39:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of pull-images.\u001b[0m\n\u001b[32m2026-03-21 00:39:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c302b01c-13af-43e8-aa09-d4d9076680ca (pull-images) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 00:39:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c302b01c-13af-43e8-aa09-d4d9076680ca is running in background. No more output. Check ARA for logs.\u001b[0m\n\u001b[32m2026-03-21 00:39:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play wipe-partitions in environment custom\u001b[0m\n\u001b[32m2026-03-21 00:39:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of wipe-partitions.\u001b[0m\n\u001b[32m2026-03-21 00:39:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 619848fa-4eb2-47d1-a281-4cc0601421ae (wipe-partitions) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 00:39:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 619848fa-4eb2-47d1-a281-4cc0601421ae (wipe-partitions) has been started and output is visible here.\u001b[0m\n\nPLAY [Wipe partitions] *********************************************************\n\nTASK [Find all logical devices owned by UID 167] *******************************\nSaturday 21 March 2026  00:40:00 +0000 (0:00:00.166)       0:00:00.166 ******** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [Remove all rook related logical devices] *********************************\nSaturday 21 March 2026  00:40:01 +0000 (0:00:01.194)       0:00:01.360 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Find all logical devices with prefix ceph] *******************************\nSaturday 21 March 2026  00:40:01 +0000 (0:00:00.218)       0:00:01.579 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Remove all ceph related logical devices] *********************************\nSaturday 21 March 2026  00:40:02 +0000 (0:00:00.531)       0:00:02.110 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Check device availability] ***********************************************\nSaturday 21 March 2026  00:40:02 +0000 (0:00:00.236)       0:00:02.347 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdd)\u001b[0m\n\nTASK [Wipe partitions with wipefs] *********************************************\nSaturday 21 March 2026  00:40:04 +0000 (0:00:01.439)       0:00:03.787 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdb)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdb)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdb)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdc)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdc)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdc)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdd)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdd)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdd)\u001b[0m\n\nTASK [Overwrite first 32M with zeros] ******************************************\nSaturday 21 March 2026  00:40:05 +0000 (0:00:01.363)       0:00:05.150 ******** \n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdd)\u001b[0m\n\nTASK [Reload udev rules] *******************************************************\nSaturday 21 March 2026  00:40:08 +0000 (0:00:03.187)       0:00:08.337 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Request device events from the kernel] ***********************************\nSaturday 21 March 2026  00:40:09 +0000 (0:00:00.599)       0:00:08.937 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:40:09 +0000 (0:00:00.601)       0:00:09.538 ******** \n=============================================================================== \nOverwrite first 32M with zeros ------------------------------------------ 3.19s\nCheck device availability ----------------------------------------------- 1.44s\nWipe partitions with wipefs --------------------------------------------- 1.36s\nFind all logical devices owned by UID 167 ------------------------------- 1.19s\nRequest device events from the kernel ----------------------------------- 0.60s\nReload udev rules ------------------------------------------------------- 0.60s\nFind all logical devices with prefix ceph ------------------------------- 0.53s\nRemove all ceph related logical devices --------------------------------- 0.24s\nRemove all rook related logical devices --------------------------------- 0.22s\n\u001b[32m2026-03-21 00:40:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of facts.\u001b[0m\n\u001b[32m2026-03-21 00:40:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ed9c7fdc-dbc7-44ff-b976-11256150f150 (facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 00:40:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task ed9c7fdc-dbc7-44ff-b976-11256150f150 (facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role facts] ********************************************************\n\nTASK [osism.commons.facts : Create custom facts directory] *********************\nSaturday 21 March 2026  00:40:25 +0000 (0:00:00.325)       0:00:00.325 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.facts : Copy fact files] ***********************************\nSaturday 21 March 2026  00:40:26 +0000 (0:00:01.268)       0:00:01.593 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nSaturday 21 March 2026  00:40:27 +0000 (0:00:01.035)       0:00:02.629 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts if using --limit] *****************************\n\nTASK [Gather facts for all hosts] **********************************************\nSaturday 21 March 2026  00:40:32 +0000 (0:00:05.137)       0:00:07.767 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:40:32 +0000 (0:00:00.450)       0:00:08.217 ******** \n=============================================================================== \nGathers facts about hosts ----------------------------------------------- 5.14s\nosism.commons.facts : Create custom facts directory --------------------- 1.27s\nosism.commons.facts : Copy fact files ----------------------------------- 1.04s\nGather facts for all hosts ---------------------------------------------- 0.45s\n\u001b[32m2026-03-21 00:40:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of ceph-configure-lvm-volumes.\u001b[0m\n\u001b[32m2026-03-21 00:40:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4f593d2c-7953-4545-b013-3439dbcf4afd (ceph-configure-lvm-volumes) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 00:40:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 4f593d2c-7953-4545-b013-3439dbcf4afd (ceph-configure-lvm-volumes) has been started and output is visible here.\u001b[0m\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Ceph configure LVM] ******************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nSaturday 21 March 2026  00:40:38 +0000 (0:00:00.258)       0:00:00.258 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nSaturday 21 March 2026  00:40:39 +0000 (0:00:00.218)       0:00:00.476 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:39 +0000 (0:00:00.205)       0:00:00.682 ******** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:39 +0000 (0:00:00.317)       0:00:00.999 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:39 +0000 (0:00:00.370)       0:00:01.370 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:40 +0000 (0:00:00.164)       0:00:01.534 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:40 +0000 (0:00:00.172)       0:00:01.707 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:40 +0000 (0:00:00.174)       0:00:01.881 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:40 +0000 (0:00:00.165)       0:00:02.047 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:40 +0000 (0:00:00.197)       0:00:02.244 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:41 +0000 (0:00:00.181)       0:00:02.426 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:41 +0000 (0:00:00.169)       0:00:02.595 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:41 +0000 (0:00:00.392)       0:00:02.988 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_f08311eb-e2ad-4eb7-8da9-ac6d899a2409)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_f08311eb-e2ad-4eb7-8da9-ac6d899a2409)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:41 +0000 (0:00:00.389)       0:00:03.377 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_e03b171b-b829-41a0-a0b6-cc46d85f013d)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_e03b171b-b829-41a0-a0b6-cc46d85f013d)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:42 +0000 (0:00:00.516)       0:00:03.893 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_9fa3c153-0e2a-45ef-8e48-e0095eb937ad)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_9fa3c153-0e2a-45ef-8e48-e0095eb937ad)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:42 +0000 (0:00:00.525)       0:00:04.419 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:43 +0000 (0:00:00.578)       0:00:04.997 ******** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:43 +0000 (0:00:00.329)       0:00:05.327 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:44 +0000 (0:00:00.173)       0:00:05.501 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:44 +0000 (0:00:00.207)       0:00:05.709 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:44 +0000 (0:00:00.187)       0:00:05.896 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:44 +0000 (0:00:00.186)       0:00:06.083 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:44 +0000 (0:00:00.174)       0:00:06.258 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:45 +0000 (0:00:00.175)       0:00:06.434 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:45 +0000 (0:00:00.174)       0:00:06.608 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:45 +0000 (0:00:00.173)       0:00:06.782 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:46 +0000 (0:00:00.797)       0:00:07.579 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:46 +0000 (0:00:00.180)       0:00:07.760 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:46 +0000 (0:00:00.162)       0:00:07.922 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:46 +0000 (0:00:00.194)       0:00:08.116 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Set UUIDs for OSD VGs/LVs] ***********************************************\nSaturday 21 March 2026  00:40:46 +0000 (0:00:00.212)       0:00:08.329 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': None})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': None})\u001b[0m\n\nTASK [Generate WAL VG names] ***************************************************\nSaturday 21 March 2026  00:40:47 +0000 (0:00:00.167)       0:00:08.497 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Generate DB VG names] ****************************************************\nSaturday 21 March 2026  00:40:47 +0000 (0:00:00.138)       0:00:08.635 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Generate shared DB/WAL VG names] *****************************************\nSaturday 21 March 2026  00:40:47 +0000 (0:00:00.125)       0:00:08.760 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Define lvm_volumes structures] *******************************************\nSaturday 21 March 2026  00:40:47 +0000 (0:00:00.131)       0:00:08.892 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block only)] *****************************\nSaturday 21 March 2026  00:40:47 +0000 (0:00:00.125)       0:00:09.017 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '64dd100a-6bcc-5763-b557-0732c79bfd55'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '5368e13a-d21f-5912-91bd-da537815d9e8'}})\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db)] *****************************\nSaturday 21 March 2026  00:40:47 +0000 (0:00:00.163)       0:00:09.181 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '64dd100a-6bcc-5763-b557-0732c79bfd55'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '5368e13a-d21f-5912-91bd-da537815d9e8'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + wal)] ****************************\nSaturday 21 March 2026  00:40:47 +0000 (0:00:00.148)       0:00:09.330 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '64dd100a-6bcc-5763-b557-0732c79bfd55'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '5368e13a-d21f-5912-91bd-da537815d9e8'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db + wal)] ***********************\nSaturday 21 March 2026  00:40:48 +0000 (0:00:00.337)       0:00:09.667 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '64dd100a-6bcc-5763-b557-0732c79bfd55'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '5368e13a-d21f-5912-91bd-da537815d9e8'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Compile lvm_volumes] *****************************************************\nSaturday 21 March 2026  00:40:48 +0000 (0:00:00.156)       0:00:09.824 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Set OSD devices config data] *********************************************\nSaturday 21 March 2026  00:40:48 +0000 (0:00:00.139)       0:00:09.963 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Set DB devices config data] **********************************************\nSaturday 21 March 2026  00:40:48 +0000 (0:00:00.145)       0:00:10.108 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Set WAL devices config data] *********************************************\nSaturday 21 March 2026  00:40:48 +0000 (0:00:00.119)       0:00:10.228 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Set DB+WAL devices config data] ******************************************\nSaturday 21 March 2026  00:40:48 +0000 (0:00:00.130)       0:00:10.358 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print ceph_osd_devices] **************************************************\nSaturday 21 March 2026  00:40:49 +0000 (0:00:00.139)       0:00:10.498 ******** \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;32m        \"sdb\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"64dd100a-6bcc-5763-b557-0732c79bfd55\"\u001b[0m\n\u001b[0;32m        },\u001b[0m\n\u001b[0;32m        \"sdc\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"5368e13a-d21f-5912-91bd-da537815d9e8\"\u001b[0m\n\u001b[0;32m        }\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print WAL devices] *******************************************************\nSaturday 21 March 2026  00:40:49 +0000 (0:00:00.141)       0:00:10.640 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print DB devices] ********************************************************\nSaturday 21 March 2026  00:40:49 +0000 (0:00:00.130)       0:00:10.770 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print shared DB/WAL devices] *********************************************\nSaturday 21 March 2026  00:40:49 +0000 (0:00:00.125)       0:00:10.896 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print configuration data] ************************************************\nSaturday 21 March 2026  00:40:49 +0000 (0:00:00.124)       0:00:11.020 ******** \n\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m\n\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m\n\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;33m            \"sdb\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"64dd100a-6bcc-5763-b557-0732c79bfd55\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            \"sdc\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"5368e13a-d21f-5912-91bd-da537815d9e8\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        },\u001b[0m\n\u001b[0;33m        \"lvm_volumes\": [\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-64dd100a-6bcc-5763-b557-0732c79bfd55\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-5368e13a-d21f-5912-91bd-da537815d9e8\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-5368e13a-d21f-5912-91bd-da537815d9e8\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        ]\u001b[0m\n\u001b[0;33m    }\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nRUNNING HANDLER [Write configuration file] *************************************\nSaturday 21 March 2026  00:40:49 +0000 (0:00:00.207)       0:00:11.228 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nPLAY [Ceph configure LVM] ******************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nSaturday 21 March 2026  00:40:52 +0000 (0:00:02.229)       0:00:13.458 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nSaturday 21 March 2026  00:40:52 +0000 (0:00:00.249)       0:00:13.708 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:52 +0000 (0:00:00.242)       0:00:13.950 ******** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:52 +0000 (0:00:00.351)       0:00:14.302 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:53 +0000 (0:00:00.197)       0:00:14.500 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:53 +0000 (0:00:00.192)       0:00:14.693 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:53 +0000 (0:00:00.205)       0:00:14.899 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:53 +0000 (0:00:00.189)       0:00:15.088 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:54 +0000 (0:00:00.572)       0:00:15.661 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:54 +0000 (0:00:00.196)       0:00:15.858 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:54 +0000 (0:00:00.206)       0:00:16.064 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:54 +0000 (0:00:00.186)       0:00:16.250 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:55 +0000 (0:00:00.421)       0:00:16.672 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_4e74a544-8b3a-4a05-98a1-b7cc13f13c18)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_4e74a544-8b3a-4a05-98a1-b7cc13f13c18)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:55 +0000 (0:00:00.418)       0:00:17.090 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_666e73fe-87f5-4f65-9ff2-95a038c8b18d)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_666e73fe-87f5-4f65-9ff2-95a038c8b18d)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:56 +0000 (0:00:00.427)       0:00:17.517 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_8ae4aabb-0174-4999-a1d0-4bf78cbb2965)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_8ae4aabb-0174-4999-a1d0-4bf78cbb2965)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:40:56 +0000 (0:00:00.418)       0:00:17.936 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:56 +0000 (0:00:00.307)       0:00:18.244 ******** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:57 +0000 (0:00:00.372)       0:00:18.616 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:57 +0000 (0:00:00.185)       0:00:18.802 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:57 +0000 (0:00:00.560)       0:00:19.363 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:58 +0000 (0:00:00.171)       0:00:19.535 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:58 +0000 (0:00:00.169)       0:00:19.705 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:58 +0000 (0:00:00.230)       0:00:19.935 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:58 +0000 (0:00:00.213)       0:00:20.149 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:58 +0000 (0:00:00.166)       0:00:20.315 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:59 +0000 (0:00:00.193)       0:00:20.508 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:59 +0000 (0:00:00.517)       0:00:21.026 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:59 +0000 (0:00:00.146)       0:00:21.173 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:40:59 +0000 (0:00:00.150)       0:00:21.323 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:41:00 +0000 (0:00:00.143)       0:00:21.466 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Set UUIDs for OSD VGs/LVs] ***********************************************\nSaturday 21 March 2026  00:41:00 +0000 (0:00:00.155)       0:00:21.622 ******** \n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': None})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': None})\u001b[0m\n\nTASK [Generate WAL VG names] ***************************************************\nSaturday 21 March 2026  00:41:00 +0000 (0:00:00.321)       0:00:21.944 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Generate DB VG names] ****************************************************\nSaturday 21 March 2026  00:41:00 +0000 (0:00:00.094)       0:00:22.038 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Generate shared DB/WAL VG names] *****************************************\nSaturday 21 March 2026  00:41:00 +0000 (0:00:00.110)       0:00:22.148 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Define lvm_volumes structures] *******************************************\nSaturday 21 March 2026  00:41:00 +0000 (0:00:00.097)       0:00:22.246 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block only)] *****************************\nSaturday 21 March 2026  00:41:00 +0000 (0:00:00.097)       0:00:22.343 ******** \n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '45b09425-beda-50b4-a83b-1fc50b5adca4'}})\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db)] *****************************\nSaturday 21 March 2026  00:41:01 +0000 (0:00:00.119)       0:00:22.463 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '45b09425-beda-50b4-a83b-1fc50b5adca4'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + wal)] ****************************\nSaturday 21 March 2026  00:41:01 +0000 (0:00:00.112)       0:00:22.576 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '45b09425-beda-50b4-a83b-1fc50b5adca4'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db + wal)] ***********************\nSaturday 21 March 2026  00:41:01 +0000 (0:00:00.111)       0:00:22.687 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '45b09425-beda-50b4-a83b-1fc50b5adca4'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Compile lvm_volumes] *****************************************************\nSaturday 21 March 2026  00:41:01 +0000 (0:00:00.131)       0:00:22.819 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Set OSD devices config data] *********************************************\nSaturday 21 March 2026  00:41:01 +0000 (0:00:00.202)       0:00:23.021 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Set DB devices config data] **********************************************\nSaturday 21 March 2026  00:41:01 +0000 (0:00:00.184)       0:00:23.206 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Set WAL devices config data] *********************************************\nSaturday 21 March 2026  00:41:01 +0000 (0:00:00.101)       0:00:23.308 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Set DB+WAL devices config data] ******************************************\nSaturday 21 March 2026  00:41:02 +0000 (0:00:00.223)       0:00:23.532 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print ceph_osd_devices] **************************************************\nSaturday 21 March 2026  00:41:02 +0000 (0:00:00.114)       0:00:23.646 ******** \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;32m        \"sdb\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"ed04aa0b-4f5a-523b-a5d2-05072a1b260d\"\u001b[0m\n\u001b[0;32m        },\u001b[0m\n\u001b[0;32m        \"sdc\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"45b09425-beda-50b4-a83b-1fc50b5adca4\"\u001b[0m\n\u001b[0;32m        }\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print WAL devices] *******************************************************\nSaturday 21 March 2026  00:41:02 +0000 (0:00:00.115)       0:00:23.761 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print DB devices] ********************************************************\nSaturday 21 March 2026  00:41:02 +0000 (0:00:00.107)       0:00:23.868 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print shared DB/WAL devices] *********************************************\nSaturday 21 March 2026  00:41:02 +0000 (0:00:00.103)       0:00:23.972 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print configuration data] ************************************************\nSaturday 21 March 2026  00:41:02 +0000 (0:00:00.110)       0:00:24.082 ******** \n\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m\n\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m\n\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;33m            \"sdb\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"ed04aa0b-4f5a-523b-a5d2-05072a1b260d\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            \"sdc\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"45b09425-beda-50b4-a83b-1fc50b5adca4\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        },\u001b[0m\n\u001b[0;33m        \"lvm_volumes\": [\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-45b09425-beda-50b4-a83b-1fc50b5adca4\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        ]\u001b[0m\n\u001b[0;33m    }\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nRUNNING HANDLER [Write configuration file] *************************************\nSaturday 21 March 2026  00:41:02 +0000 (0:00:00.162)       0:00:24.245 ******** \n\u001b[0;33mchanged: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nPLAY [Ceph configure LVM] ******************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nSaturday 21 March 2026  00:41:03 +0000 (0:00:00.916)       0:00:25.161 ******** \n\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nSaturday 21 March 2026  00:41:04 +0000 (0:00:00.364)       0:00:25.526 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:41:04 +0000 (0:00:00.494)       0:00:26.020 ******** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:41:04 +0000 (0:00:00.287)       0:00:26.308 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:41:05 +0000 (0:00:00.185)       0:00:26.493 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:41:05 +0000 (0:00:00.174)       0:00:26.668 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:41:05 +0000 (0:00:00.196)       0:00:26.865 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:41:05 +0000 (0:00:00.188)       0:00:27.054 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:41:05 +0000 (0:00:00.161)       0:00:27.215 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:41:05 +0000 (0:00:00.156)       0:00:27.372 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:41:06 +0000 (0:00:00.190)       0:00:27.563 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:41:06 +0000 (0:00:00.187)       0:00:27.750 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:41:06 +0000 (0:00:00.528)       0:00:28.278 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_f76ae8f0-65cc-4a62-b19c-23f8a40fbde6)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_f76ae8f0-65cc-4a62-b19c-23f8a40fbde6)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:41:07 +0000 (0:00:00.672)       0:00:28.951 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_d7bb81ba-6491-4b5a-bb82-f0b49314eac0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_d7bb81ba-6491-4b5a-bb82-f0b49314eac0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:41:07 +0000 (0:00:00.415)       0:00:29.366 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_cdffe5f1-7f0d-460e-a1db-d6846d226144)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_cdffe5f1-7f0d-460e-a1db-d6846d226144)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:41:08 +0000 (0:00:00.371)       0:00:29.738 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:41:08 +0000 (0:00:00.337)       0:00:30.075 ******** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:41:08 +0000 (0:00:00.347)       0:00:30.423 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:41:09 +0000 (0:00:00.210)       0:00:30.634 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:41:09 +0000 (0:00:00.170)       0:00:30.805 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:41:09 +0000 (0:00:00.187)       0:00:30.993 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:41:09 +0000 (0:00:00.182)       0:00:31.175 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:41:09 +0000 (0:00:00.182)       0:00:31.357 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:41:10 +0000 (0:00:00.507)       0:00:31.865 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:41:10 +0000 (0:00:00.231)       0:00:32.097 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:41:10 +0000 (0:00:00.218)       0:00:32.316 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:41:11 +0000 (0:00:00.639)       0:00:32.956 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:41:11 +0000 (0:00:00.167)       0:00:33.123 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:41:11 +0000 (0:00:00.159)       0:00:33.283 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:41:12 +0000 (0:00:00.169)       0:00:33.453 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set UUIDs for OSD VGs/LVs] ***********************************************\nSaturday 21 March 2026  00:41:12 +0000 (0:00:00.167)       0:00:33.620 ******** \n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': None})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': None})\u001b[0m\n\nTASK [Generate WAL VG names] ***************************************************\nSaturday 21 March 2026  00:41:12 +0000 (0:00:00.150)       0:00:33.770 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Generate DB VG names] ****************************************************\nSaturday 21 March 2026  00:41:12 +0000 (0:00:00.127)       0:00:33.898 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Generate shared DB/WAL VG names] *****************************************\nSaturday 21 March 2026  00:41:12 +0000 (0:00:00.116)       0:00:34.014 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Define lvm_volumes structures] *******************************************\nSaturday 21 March 2026  00:41:12 +0000 (0:00:00.100)       0:00:34.115 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block only)] *****************************\nSaturday 21 March 2026  00:41:13 +0000 (0:00:00.317)       0:00:34.432 ******** \n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'ede169e1-d974-55ed-a938-c303cb89cb5b'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}})\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db)] *****************************\nSaturday 21 March 2026  00:41:13 +0000 (0:00:00.206)       0:00:34.639 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'ede169e1-d974-55ed-a938-c303cb89cb5b'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + wal)] ****************************\nSaturday 21 March 2026  00:41:13 +0000 (0:00:00.166)       0:00:34.806 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'ede169e1-d974-55ed-a938-c303cb89cb5b'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db + wal)] ***********************\nSaturday 21 March 2026  00:41:13 +0000 (0:00:00.163)       0:00:34.970 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'ede169e1-d974-55ed-a938-c303cb89cb5b'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Compile lvm_volumes] *****************************************************\nSaturday 21 March 2026  00:41:13 +0000 (0:00:00.146)       0:00:35.116 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set OSD devices config data] *********************************************\nSaturday 21 March 2026  00:41:13 +0000 (0:00:00.116)       0:00:35.233 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set DB devices config data] **********************************************\nSaturday 21 March 2026  00:41:13 +0000 (0:00:00.115)       0:00:35.348 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set WAL devices config data] *********************************************\nSaturday 21 March 2026  00:41:14 +0000 (0:00:00.103)       0:00:35.452 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set DB+WAL devices config data] ******************************************\nSaturday 21 March 2026  00:41:14 +0000 (0:00:00.119)       0:00:35.571 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print ceph_osd_devices] **************************************************\nSaturday 21 March 2026  00:41:14 +0000 (0:00:00.137)       0:00:35.709 ******** \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;32m        \"sdb\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"ede169e1-d974-55ed-a938-c303cb89cb5b\"\u001b[0m\n\u001b[0;32m        },\u001b[0m\n\u001b[0;32m        \"sdc\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"23a81e86-5308-55a4-b4d4-ad2cd0b09f4a\"\u001b[0m\n\u001b[0;32m        }\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print WAL devices] *******************************************************\nSaturday 21 March 2026  00:41:14 +0000 (0:00:00.112)       0:00:35.821 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print DB devices] ********************************************************\nSaturday 21 March 2026  00:41:14 +0000 (0:00:00.095)       0:00:35.917 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print shared DB/WAL devices] *********************************************\nSaturday 21 March 2026  00:41:14 +0000 (0:00:00.266)       0:00:36.183 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print configuration data] ************************************************\nSaturday 21 March 2026  00:41:14 +0000 (0:00:00.116)       0:00:36.300 ******** \n\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m\n\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m\n\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;33m            \"sdb\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"ede169e1-d974-55ed-a938-c303cb89cb5b\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            \"sdc\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"23a81e86-5308-55a4-b4d4-ad2cd0b09f4a\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        },\u001b[0m\n\u001b[0;33m        \"lvm_volumes\": [\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-ede169e1-d974-55ed-a938-c303cb89cb5b\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        ]\u001b[0m\n\u001b[0;33m    }\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nRUNNING HANDLER [Write configuration file] *************************************\nSaturday 21 March 2026  00:41:15 +0000 (0:00:00.186)       0:00:36.486 ******** \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:41:15 +0000 (0:00:00.786)       0:00:37.273 ******** \n=============================================================================== \nWrite configuration file ------------------------------------------------ 3.93s\nAdd known partitions to the list of available block devices ------------- 1.05s\nAdd known links to the list of available block devices ------------------ 0.96s\nGet initial list of available block devices ----------------------------- 0.94s\nGet extra vars for Ceph configuration ----------------------------------- 0.83s\nAdd known partitions to the list of available block devices ------------- 0.80s\nAdd known links to the list of available block devices ------------------ 0.67s\nAdd known partitions to the list of available block devices ------------- 0.64s\nSet UUIDs for OSD VGs/LVs ----------------------------------------------- 0.64s\nGenerate lvm_volumes structure (block + wal) ---------------------------- 0.61s\nAdd known links to the list of available block devices ------------------ 0.58s\nAdd known links to the list of available block devices ------------------ 0.57s\nAdd known partitions to the list of available block devices ------------- 0.56s\nPrint configuration data ------------------------------------------------ 0.56s\nDefine lvm_volumes structures ------------------------------------------- 0.54s\nAdd known links to the list of available block devices ------------------ 0.53s\nAdd known links to the list of available block devices ------------------ 0.53s\nAdd known partitions to the list of available block devices ------------- 0.52s\nAdd known links to the list of available block devices ------------------ 0.52s\nAdd known partitions to the list of available block devices ------------- 0.51s\n\u001b[32m2026-03-21 00:41:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 069d0a36-c9b8-49ce-bd82-37b1b73f47ae (sync inventory) is running in background. Output coming soon.\u001b[0m\n\u001b[32m2026-03-21 00:41:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting group_vars file reorganization\u001b[0m\n\u001b[32m2026-03-21 00:41:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMoved 0 file(s) to their respective directories\u001b[0m\n\u001b[32m2026-03-21 00:41:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGroup_vars file reorganization completed\u001b[0m\n\u001b[32m2026-03-21 00:41:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting variable preparation from inventory\u001b[0m\n\u001b[32m2026-03-21 00:41:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-kolla-ceph-rgw-hosts.yml with ceph_rgw_hosts\u001b[0m\n\u001b[32m2026-03-21 00:41:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-infrastructure-cephclient-mons.yml with cephclient_mons\u001b[0m\n\u001b[32m2026-03-21 00:41:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-ceph-cluster-fsid.yml with ceph_cluster_fsid\u001b[0m\n\u001b[32m2026-03-21 00:41:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m3 file(s) written, 6 host(s) processed\u001b[0m\n\u001b[32m2026-03-21 00:41:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mVariable preparation completed\u001b[0m\n\u001b[32m2026-03-21 00:41:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting inventory overwrite handling\u001b[0m\n\u001b[32m2026-03-21 00:41:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mHandling group overwrites in 99-overwrite\u001b[0m\n\u001b[32m2026-03-21 00:41:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group frr:children from 60-generic\u001b[0m\n\u001b[32m2026-03-21 00:41:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group netbird:children from 50-infrastructure\u001b[0m\n\u001b[32m2026-03-21 00:41:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group ceph-mds from 50-ceph\u001b[0m\n\u001b[32m2026-03-21 00:41:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group ceph-rgw from 50-ceph\u001b[0m\n\u001b[32m2026-03-21 00:41:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mHandling group overwrites in 20-roles\u001b[0m\n\u001b[32m2026-03-21 00:41:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group k3s_node from 50-infrastructure\u001b[0m\n\u001b[32m2026-03-21 00:41:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoved 5 group(s) in total\u001b[0m\n\u001b[32m2026-03-21 00:41:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mInventory overwrite handling completed\u001b[0m\n\u001b[32m2026-03-21 00:41:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting merge of inventory files\u001b[0m\n\u001b[32m2026-03-21 00:41:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mInventory files merged successfully\u001b[0m\n\u001b[32m2026-03-21 00:41:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGenerating minified hosts file\u001b[0m\n\u001b[32m2026-03-21 00:41:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSuccessfully wrote minified hosts file to /inventory.merge/hosts-minified.yml\u001b[0m\n\u001b[32m2026-03-21 00:41:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSuccessfully wrote fast inventory to /inventory.merge/fast/hosts.json\u001b[0m\n\u001b[32m2026-03-21 00:41:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGenerating ClusterShell configuration from Ansible inventory\u001b[0m\n\u001b[32m2026-03-21 00:42:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSuccessfully wrote ClusterShell configuration\u001b[0m\n[master d79765b] 2026-03-21-00-42\n 5 files changed, 75 insertions(+), 10 deletions(-)\n create mode 100644 fast/host_vars/testbed-node-3/ceph-lvm-configuration.yml\n create mode 100644 fast/host_vars/testbed-node-4/ceph-lvm-configuration.yml\n create mode 100644 fast/host_vars/testbed-node-5/ceph-lvm-configuration.yml\n\u001b[32m2026-03-21 00:42:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of ceph-create-lvm-devices.\u001b[0m\n\u001b[32m2026-03-21 00:42:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c172153-90a3-4f8f-a4c7-2223f84e593f (ceph-create-lvm-devices) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 00:42:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 0c172153-90a3-4f8f-a4c7-2223f84e593f (ceph-create-lvm-devices) has been started and output is visible here.\u001b[0m\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Ceph create LVM devices] *************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nSaturday 21 March 2026  00:42:14 +0000 (0:00:00.290)       0:00:00.290 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nSaturday 21 March 2026  00:42:14 +0000 (0:00:00.236)       0:00:00.527 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:14 +0000 (0:00:00.215)       0:00:00.742 ******** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:15 +0000 (0:00:00.380)       0:00:01.123 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:15 +0000 (0:00:00.458)       0:00:01.582 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:15 +0000 (0:00:00.198)       0:00:01.780 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:16 +0000 (0:00:00.184)       0:00:01.965 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:16 +0000 (0:00:00.190)       0:00:02.155 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:16 +0000 (0:00:00.191)       0:00:02.347 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:16 +0000 (0:00:00.222)       0:00:02.569 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:16 +0000 (0:00:00.204)       0:00:02.773 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:17 +0000 (0:00:00.188)       0:00:02.962 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:17 +0000 (0:00:00.423)       0:00:03.385 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_f08311eb-e2ad-4eb7-8da9-ac6d899a2409)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_f08311eb-e2ad-4eb7-8da9-ac6d899a2409)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:17 +0000 (0:00:00.417)       0:00:03.802 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_e03b171b-b829-41a0-a0b6-cc46d85f013d)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_e03b171b-b829-41a0-a0b6-cc46d85f013d)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:18 +0000 (0:00:00.650)       0:00:04.453 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_9fa3c153-0e2a-45ef-8e48-e0095eb937ad)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_9fa3c153-0e2a-45ef-8e48-e0095eb937ad)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:19 +0000 (0:00:00.625)       0:00:05.079 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:20 +0000 (0:00:00.821)       0:00:05.901 ******** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:20 +0000 (0:00:00.496)       0:00:06.397 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:20 +0000 (0:00:00.218)       0:00:06.616 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:21 +0000 (0:00:00.226)       0:00:06.843 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:21 +0000 (0:00:00.218)       0:00:07.062 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:21 +0000 (0:00:00.219)       0:00:07.281 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:21 +0000 (0:00:00.207)       0:00:07.489 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:21 +0000 (0:00:00.229)       0:00:07.718 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:22 +0000 (0:00:00.228)       0:00:07.947 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:22 +0000 (0:00:00.214)       0:00:08.162 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:23 +0000 (0:00:01.366)       0:00:09.529 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:23 +0000 (0:00:00.212)       0:00:09.741 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:24 +0000 (0:00:00.212)       0:00:09.954 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:24 +0000 (0:00:00.214)       0:00:10.168 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Check whether ceph_db_wal_devices is used exclusively] *******************\nSaturday 21 March 2026  00:42:24 +0000 (0:00:00.232)       0:00:10.401 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************\nSaturday 21 March 2026  00:42:24 +0000 (0:00:00.140)       0:00:10.541 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '64dd100a-6bcc-5763-b557-0732c79bfd55'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '5368e13a-d21f-5912-91bd-da537815d9e8'}})\u001b[0m\n\nTASK [Create block VGs] ********************************************************\nSaturday 21 March 2026  00:42:24 +0000 (0:00:00.224)       0:00:10.765 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'})\u001b[0m\n\nTASK [Print 'Create block VGs'] ************************************************\nSaturday 21 March 2026  00:42:26 +0000 (0:00:02.026)       0:00:12.792 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create block LVs] ********************************************************\nSaturday 21 March 2026  00:42:27 +0000 (0:00:00.159)       0:00:12.951 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'})\u001b[0m\n\nTASK [Print 'Create block LVs'] ************************************************\nSaturday 21 March 2026  00:42:28 +0000 (0:00:01.455)       0:00:14.406 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create DB VGs] ***********************************************************\nSaturday 21 March 2026  00:42:28 +0000 (0:00:00.179)       0:00:14.586 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create DB VGs'] ***************************************************\nSaturday 21 March 2026  00:42:28 +0000 (0:00:00.149)       0:00:14.736 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create WAL VGs] **********************************************************\nSaturday 21 March 2026  00:42:29 +0000 (0:00:00.347)       0:00:15.083 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create WAL VGs'] **************************************************\nSaturday 21 March 2026  00:42:29 +0000 (0:00:00.156)       0:00:15.240 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create DB+WAL VGs] *******************************************************\nSaturday 21 March 2026  00:42:29 +0000 (0:00:00.173)       0:00:15.413 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create DB+WAL VGs'] ***********************************************\nSaturday 21 March 2026  00:42:29 +0000 (0:00:00.179)       0:00:15.592 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Prepare variables for OSD count check] ***********************************\nSaturday 21 March 2026  00:42:29 +0000 (0:00:00.189)       0:00:15.782 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************\nSaturday 21 March 2026  00:42:30 +0000 (0:00:00.138)       0:00:15.920 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************\nSaturday 21 March 2026  00:42:30 +0000 (0:00:00.210)       0:00:16.131 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************\nSaturday 21 March 2026  00:42:30 +0000 (0:00:00.173)       0:00:16.305 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************\nSaturday 21 March 2026  00:42:30 +0000 (0:00:00.175)       0:00:16.480 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************\nSaturday 21 March 2026  00:42:30 +0000 (0:00:00.146)       0:00:16.626 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************\nSaturday 21 March 2026  00:42:30 +0000 (0:00:00.174)       0:00:16.800 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print number of OSDs wanted per DB VG] ***********************************\nSaturday 21 March 2026  00:42:31 +0000 (0:00:00.153)       0:00:16.954 ******** \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per WAL VG] **********************************\nSaturday 21 March 2026  00:42:31 +0000 (0:00:00.510)       0:00:17.464 ******** \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per DB+WAL VG] *******************************\nSaturday 21 March 2026  00:42:31 +0000 (0:00:00.154)       0:00:17.619 ******** \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Gather DB VGs with total and available size in bytes] ********************\nSaturday 21 March 2026  00:42:31 +0000 (0:00:00.148)       0:00:17.768 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Gather WAL VGs with total and available size in bytes] *******************\nSaturday 21 March 2026  00:42:32 +0000 (0:00:00.746)       0:00:18.515 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Gather DB+WAL VGs with total and available size in bytes] ****************\nSaturday 21 March 2026  00:42:33 +0000 (0:00:00.490)       0:00:19.005 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************\nSaturday 21 March 2026  00:42:33 +0000 (0:00:00.543)       0:00:19.548 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Calculate VG sizes (without buffer)] *************************************\nSaturday 21 March 2026  00:42:33 +0000 (0:00:00.136)       0:00:19.685 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate VG sizes (with buffer)] ****************************************\nSaturday 21 March 2026  00:42:33 +0000 (0:00:00.120)       0:00:19.805 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print LVM VGs report data] ***********************************************\nSaturday 21 March 2026  00:42:34 +0000 (0:00:00.105)       0:00:19.910 ******** \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"vgs_report\": {\u001b[0m\n\u001b[0;32m        \"vg\": []\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print LVM VG sizes] ******************************************************\nSaturday 21 March 2026  00:42:34 +0000 (0:00:00.134)       0:00:20.045 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_db_devices] ************************\nSaturday 21 March 2026  00:42:34 +0000 (0:00:00.149)       0:00:20.194 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_db_devices] ****************************\nSaturday 21 March 2026  00:42:34 +0000 (0:00:00.133)       0:00:20.328 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if size of DB LVs on ceph_db_devices > available] *******************\nSaturday 21 March 2026  00:42:34 +0000 (0:00:00.358)       0:00:20.686 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_wal_devices] ***********************\nSaturday 21 March 2026  00:42:35 +0000 (0:00:00.142)       0:00:20.828 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_wal_devices] ***************************\nSaturday 21 March 2026  00:42:35 +0000 (0:00:00.132)       0:00:20.961 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************\nSaturday 21 March 2026  00:42:35 +0000 (0:00:00.132)       0:00:21.094 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************\nSaturday 21 March 2026  00:42:35 +0000 (0:00:00.134)       0:00:21.229 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************\nSaturday 21 March 2026  00:42:35 +0000 (0:00:00.135)       0:00:21.364 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************\nSaturday 21 March 2026  00:42:35 +0000 (0:00:00.141)       0:00:21.506 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************\nSaturday 21 March 2026  00:42:35 +0000 (0:00:00.138)       0:00:21.644 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********\nSaturday 21 March 2026  00:42:35 +0000 (0:00:00.127)       0:00:21.772 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************\nSaturday 21 March 2026  00:42:36 +0000 (0:00:00.131)       0:00:21.904 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************\nSaturday 21 March 2026  00:42:36 +0000 (0:00:00.126)       0:00:22.030 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_devices] ***************************************\nSaturday 21 March 2026  00:42:36 +0000 (0:00:00.139)       0:00:22.169 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_devices'] *******************************\nSaturday 21 March 2026  00:42:36 +0000 (0:00:00.188)       0:00:22.357 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create WAL LVs for ceph_wal_devices] *************************************\nSaturday 21 March 2026  00:42:36 +0000 (0:00:00.386)       0:00:22.744 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************\nSaturday 21 March 2026  00:42:37 +0000 (0:00:00.157)       0:00:22.902 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create WAL LVs for ceph_db_wal_devices] **********************************\nSaturday 21 March 2026  00:42:37 +0000 (0:00:00.173)       0:00:23.075 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************\nSaturday 21 March 2026  00:42:37 +0000 (0:00:00.157)       0:00:23.233 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_wal_devices] ***********************************\nSaturday 21 March 2026  00:42:37 +0000 (0:00:00.154)       0:00:23.388 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************\nSaturday 21 March 2026  00:42:37 +0000 (0:00:00.150)       0:00:23.538 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Get list of Ceph LVs with associated VGs] ********************************\nSaturday 21 March 2026  00:42:37 +0000 (0:00:00.173)       0:00:23.712 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Get list of Ceph PVs with associated VGs] ********************************\nSaturday 21 March 2026  00:42:38 +0000 (0:00:00.500)       0:00:24.213 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************\nSaturday 21 March 2026  00:42:38 +0000 (0:00:00.533)       0:00:24.746 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Create list of VG/LV names] **********************************************\nSaturday 21 March 2026  00:42:39 +0000 (0:00:00.151)       0:00:24.898 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'lv_name': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'vg_name': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'lv_name': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'vg_name': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'})\u001b[0m\n\nTASK [Fail if block LV defined in lvm_volumes is missing] **********************\nSaturday 21 March 2026  00:42:39 +0000 (0:00:00.183)       0:00:25.082 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if DB LV defined in lvm_volumes is missing] *************************\nSaturday 21 March 2026  00:42:39 +0000 (0:00:00.148)       0:00:25.230 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if WAL LV defined in lvm_volumes is missing] ************************\nSaturday 21 March 2026  00:42:39 +0000 (0:00:00.368)       0:00:25.599 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print LVM report data] ***************************************************\nSaturday 21 March 2026  00:42:39 +0000 (0:00:00.171)       0:00:25.771 ******** \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"lvm_report\": {\u001b[0m\n\u001b[0;32m        \"lv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-5368e13a-d21f-5912-91bd-da537815d9e8\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-5368e13a-d21f-5912-91bd-da537815d9e8\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-64dd100a-6bcc-5763-b557-0732c79bfd55\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ],\u001b[0m\n\u001b[0;32m        \"pv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-64dd100a-6bcc-5763-b557-0732c79bfd55\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-5368e13a-d21f-5912-91bd-da537815d9e8\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ]\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY [Ceph create LVM devices] *************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nSaturday 21 March 2026  00:42:40 +0000 (0:00:00.304)       0:00:26.076 ******** \n\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nSaturday 21 March 2026  00:42:40 +0000 (0:00:00.262)       0:00:26.338 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:40 +0000 (0:00:00.229)       0:00:26.568 ******** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:41 +0000 (0:00:00.387)       0:00:26.956 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:41 +0000 (0:00:00.196)       0:00:27.152 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:41 +0000 (0:00:00.215)       0:00:27.368 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:41 +0000 (0:00:00.226)       0:00:27.595 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:42 +0000 (0:00:00.641)       0:00:28.237 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:42 +0000 (0:00:00.198)       0:00:28.436 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:42 +0000 (0:00:00.222)       0:00:28.658 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:43 +0000 (0:00:00.221)       0:00:28.880 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:43 +0000 (0:00:00.226)       0:00:29.107 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:43 +0000 (0:00:00.467)       0:00:29.574 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_4e74a544-8b3a-4a05-98a1-b7cc13f13c18)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_4e74a544-8b3a-4a05-98a1-b7cc13f13c18)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:44 +0000 (0:00:00.418)       0:00:29.993 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_666e73fe-87f5-4f65-9ff2-95a038c8b18d)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_666e73fe-87f5-4f65-9ff2-95a038c8b18d)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:44 +0000 (0:00:00.384)       0:00:30.378 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_8ae4aabb-0174-4999-a1d0-4bf78cbb2965)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_8ae4aabb-0174-4999-a1d0-4bf78cbb2965)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:42:44 +0000 (0:00:00.390)       0:00:30.769 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:45 +0000 (0:00:00.303)       0:00:31.072 ******** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:45 +0000 (0:00:00.490)       0:00:31.563 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:45 +0000 (0:00:00.167)       0:00:31.730 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:46 +0000 (0:00:00.205)       0:00:31.936 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:46 +0000 (0:00:00.185)       0:00:32.121 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:46 +0000 (0:00:00.187)       0:00:32.309 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:46 +0000 (0:00:00.175)       0:00:32.485 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:46 +0000 (0:00:00.183)       0:00:32.669 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:47 +0000 (0:00:00.175)       0:00:32.844 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:47 +0000 (0:00:00.166)       0:00:33.011 ******** \n\u001b[0;32mok: [testbed-node-4] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:47 +0000 (0:00:00.744)       0:00:33.755 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:48 +0000 (0:00:00.179)       0:00:33.935 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:48 +0000 (0:00:00.191)       0:00:34.127 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:42:48 +0000 (0:00:00.470)       0:00:34.597 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Check whether ceph_db_wal_devices is used exclusively] *******************\nSaturday 21 March 2026  00:42:48 +0000 (0:00:00.186)       0:00:34.783 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************\nSaturday 21 March 2026  00:42:49 +0000 (0:00:00.120)       0:00:34.904 ******** \n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '45b09425-beda-50b4-a83b-1fc50b5adca4'}})\u001b[0m\n\nTASK [Create block VGs] ********************************************************\nSaturday 21 March 2026  00:42:49 +0000 (0:00:00.168)       0:00:35.073 ******** \n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'})\u001b[0m\n\nTASK [Print 'Create block VGs'] ************************************************\nSaturday 21 March 2026  00:42:51 +0000 (0:00:01.843)       0:00:36.917 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create block LVs] ********************************************************\nSaturday 21 March 2026  00:42:51 +0000 (0:00:00.130)       0:00:37.047 ******** \n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'})\u001b[0m\n\nTASK [Print 'Create block LVs'] ************************************************\nSaturday 21 March 2026  00:42:52 +0000 (0:00:01.193)       0:00:38.240 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create DB VGs] ***********************************************************\nSaturday 21 March 2026  00:42:52 +0000 (0:00:00.124)       0:00:38.365 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create DB VGs'] ***************************************************\nSaturday 21 March 2026  00:42:52 +0000 (0:00:00.129)       0:00:38.494 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create WAL VGs] **********************************************************\nSaturday 21 March 2026  00:42:52 +0000 (0:00:00.135)       0:00:38.630 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create WAL VGs'] **************************************************\nSaturday 21 March 2026  00:42:52 +0000 (0:00:00.125)       0:00:38.755 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create DB+WAL VGs] *******************************************************\nSaturday 21 March 2026  00:42:53 +0000 (0:00:00.134)       0:00:38.890 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create DB+WAL VGs'] ***********************************************\nSaturday 21 March 2026  00:42:53 +0000 (0:00:00.258)       0:00:39.148 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Prepare variables for OSD count check] ***********************************\nSaturday 21 March 2026  00:42:53 +0000 (0:00:00.134)       0:00:39.283 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************\nSaturday 21 March 2026  00:42:53 +0000 (0:00:00.125)       0:00:39.408 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************\nSaturday 21 March 2026  00:42:53 +0000 (0:00:00.138)       0:00:39.546 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************\nSaturday 21 March 2026  00:42:53 +0000 (0:00:00.149)       0:00:39.695 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************\nSaturday 21 March 2026  00:42:54 +0000 (0:00:00.140)       0:00:39.836 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************\nSaturday 21 March 2026  00:42:54 +0000 (0:00:00.118)       0:00:39.955 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************\nSaturday 21 March 2026  00:42:54 +0000 (0:00:00.136)       0:00:40.091 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print number of OSDs wanted per DB VG] ***********************************\nSaturday 21 March 2026  00:42:54 +0000 (0:00:00.131)       0:00:40.222 ******** \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per WAL VG] **********************************\nSaturday 21 March 2026  00:42:54 +0000 (0:00:00.114)       0:00:40.337 ******** \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per DB+WAL VG] *******************************\nSaturday 21 March 2026  00:42:54 +0000 (0:00:00.129)       0:00:40.467 ******** \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Gather DB VGs with total and available size in bytes] ********************\nSaturday 21 March 2026  00:42:54 +0000 (0:00:00.105)       0:00:40.573 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Gather WAL VGs with total and available size in bytes] *******************\nSaturday 21 March 2026  00:42:55 +0000 (0:00:00.692)       0:00:41.266 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Gather DB+WAL VGs with total and available size in bytes] ****************\nSaturday 21 March 2026  00:42:56 +0000 (0:00:00.592)       0:00:41.858 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************\nSaturday 21 March 2026  00:42:56 +0000 (0:00:00.594)       0:00:42.452 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Calculate VG sizes (without buffer)] *************************************\nSaturday 21 March 2026  00:42:56 +0000 (0:00:00.163)       0:00:42.616 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate VG sizes (with buffer)] ****************************************\nSaturday 21 March 2026  00:42:56 +0000 (0:00:00.116)       0:00:42.732 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print LVM VGs report data] ***********************************************\nSaturday 21 March 2026  00:42:57 +0000 (0:00:00.105)       0:00:42.837 ******** \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"vgs_report\": {\u001b[0m\n\u001b[0;32m        \"vg\": []\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print LVM VG sizes] ******************************************************\nSaturday 21 March 2026  00:42:57 +0000 (0:00:00.136)       0:00:42.974 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_db_devices] ************************\nSaturday 21 March 2026  00:42:57 +0000 (0:00:00.152)       0:00:43.126 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_db_devices] ****************************\nSaturday 21 March 2026  00:42:57 +0000 (0:00:00.141)       0:00:43.268 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if size of DB LVs on ceph_db_devices > available] *******************\nSaturday 21 March 2026  00:42:57 +0000 (0:00:00.133)       0:00:43.402 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_wal_devices] ***********************\nSaturday 21 March 2026  00:42:57 +0000 (0:00:00.134)       0:00:43.537 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_wal_devices] ***************************\nSaturday 21 March 2026  00:42:57 +0000 (0:00:00.137)       0:00:43.674 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************\nSaturday 21 March 2026  00:42:58 +0000 (0:00:00.326)       0:00:44.000 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************\nSaturday 21 March 2026  00:42:58 +0000 (0:00:00.141)       0:00:44.141 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************\nSaturday 21 March 2026  00:42:58 +0000 (0:00:00.134)       0:00:44.276 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************\nSaturday 21 March 2026  00:42:58 +0000 (0:00:00.137)       0:00:44.414 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************\nSaturday 21 March 2026  00:42:58 +0000 (0:00:00.121)       0:00:44.535 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********\nSaturday 21 March 2026  00:42:58 +0000 (0:00:00.148)       0:00:44.683 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************\nSaturday 21 March 2026  00:42:59 +0000 (0:00:00.161)       0:00:44.845 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************\nSaturday 21 March 2026  00:42:59 +0000 (0:00:00.123)       0:00:44.969 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_devices] ***************************************\nSaturday 21 March 2026  00:42:59 +0000 (0:00:00.130)       0:00:45.099 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_devices'] *******************************\nSaturday 21 March 2026  00:42:59 +0000 (0:00:00.165)       0:00:45.264 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create WAL LVs for ceph_wal_devices] *************************************\nSaturday 21 March 2026  00:42:59 +0000 (0:00:00.158)       0:00:45.422 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************\nSaturday 21 March 2026  00:42:59 +0000 (0:00:00.161)       0:00:45.584 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create WAL LVs for ceph_db_wal_devices] **********************************\nSaturday 21 March 2026  00:43:00 +0000 (0:00:00.366)       0:00:45.950 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************\nSaturday 21 March 2026  00:43:00 +0000 (0:00:00.172)       0:00:46.123 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_wal_devices] ***********************************\nSaturday 21 March 2026  00:43:00 +0000 (0:00:00.149)       0:00:46.273 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************\nSaturday 21 March 2026  00:43:00 +0000 (0:00:00.134)       0:00:46.407 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Get list of Ceph LVs with associated VGs] ********************************\nSaturday 21 March 2026  00:43:00 +0000 (0:00:00.156)       0:00:46.564 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Get list of Ceph PVs with associated VGs] ********************************\nSaturday 21 March 2026  00:43:01 +0000 (0:00:00.504)       0:00:47.069 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************\nSaturday 21 March 2026  00:43:01 +0000 (0:00:00.500)       0:00:47.569 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Create list of VG/LV names] **********************************************\nSaturday 21 March 2026  00:43:01 +0000 (0:00:00.118)       0:00:47.687 ******** \n\u001b[0;32mok: [testbed-node-4] => (item={'lv_name': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'vg_name': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'lv_name': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'vg_name': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'})\u001b[0m\n\nTASK [Fail if block LV defined in lvm_volumes is missing] **********************\nSaturday 21 March 2026  00:43:02 +0000 (0:00:00.151)       0:00:47.839 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if DB LV defined in lvm_volumes is missing] *************************\nSaturday 21 March 2026  00:43:02 +0000 (0:00:00.146)       0:00:47.985 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if WAL LV defined in lvm_volumes is missing] ************************\nSaturday 21 March 2026  00:43:02 +0000 (0:00:00.147)       0:00:48.133 ******** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print LVM report data] ***************************************************\nSaturday 21 March 2026  00:43:02 +0000 (0:00:00.158)       0:00:48.291 ******** \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"lvm_report\": {\u001b[0m\n\u001b[0;32m        \"lv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-45b09425-beda-50b4-a83b-1fc50b5adca4\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ],\u001b[0m\n\u001b[0;32m        \"pv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-45b09425-beda-50b4-a83b-1fc50b5adca4\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ]\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY [Ceph create LVM devices] *************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nSaturday 21 March 2026  00:43:02 +0000 (0:00:00.401)       0:00:48.692 ******** \n\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nSaturday 21 March 2026  00:43:03 +0000 (0:00:00.217)       0:00:48.910 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:43:03 +0000 (0:00:00.226)       0:00:49.137 ******** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:43:03 +0000 (0:00:00.377)       0:00:49.515 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:43:03 +0000 (0:00:00.181)       0:00:49.696 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:43:04 +0000 (0:00:00.182)       0:00:49.878 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:43:04 +0000 (0:00:00.182)       0:00:50.061 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:43:04 +0000 (0:00:00.179)       0:00:50.241 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:43:04 +0000 (0:00:00.183)       0:00:50.424 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:43:05 +0000 (0:00:00.472)       0:00:50.896 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:43:05 +0000 (0:00:00.160)       0:00:51.057 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:43:05 +0000 (0:00:00.165)       0:00:51.223 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:43:05 +0000 (0:00:00.368)       0:00:51.591 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_f76ae8f0-65cc-4a62-b19c-23f8a40fbde6)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_f76ae8f0-65cc-4a62-b19c-23f8a40fbde6)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:43:06 +0000 (0:00:00.372)       0:00:51.964 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_d7bb81ba-6491-4b5a-bb82-f0b49314eac0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_d7bb81ba-6491-4b5a-bb82-f0b49314eac0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:43:06 +0000 (0:00:00.386)       0:00:52.351 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_cdffe5f1-7f0d-460e-a1db-d6846d226144)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_cdffe5f1-7f0d-460e-a1db-d6846d226144)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nSaturday 21 March 2026  00:43:06 +0000 (0:00:00.385)       0:00:52.736 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:43:07 +0000 (0:00:00.296)       0:00:53.033 ******** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:43:07 +0000 (0:00:00.396)       0:00:53.429 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:43:07 +0000 (0:00:00.181)       0:00:53.610 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:43:07 +0000 (0:00:00.190)       0:00:53.801 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:43:08 +0000 (0:00:00.565)       0:00:54.366 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:43:08 +0000 (0:00:00.219)       0:00:54.586 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:43:08 +0000 (0:00:00.213)       0:00:54.799 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:43:09 +0000 (0:00:00.211)       0:00:55.011 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:43:09 +0000 (0:00:00.209)       0:00:55.221 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:43:09 +0000 (0:00:00.216)       0:00:55.438 ******** \n\u001b[0;32mok: [testbed-node-5] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:43:10 +0000 (0:00:00.641)       0:00:56.080 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:43:10 +0000 (0:00:00.198)       0:00:56.279 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:43:10 +0000 (0:00:00.215)       0:00:56.494 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nSaturday 21 March 2026  00:43:10 +0000 (0:00:00.225)       0:00:56.719 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Check whether ceph_db_wal_devices is used exclusively] *******************\nSaturday 21 March 2026  00:43:11 +0000 (0:00:00.220)       0:00:56.940 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************\nSaturday 21 March 2026  00:43:11 +0000 (0:00:00.405)       0:00:57.346 ******** \n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'ede169e1-d974-55ed-a938-c303cb89cb5b'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}})\u001b[0m\n\nTASK [Create block VGs] ********************************************************\nSaturday 21 March 2026  00:43:11 +0000 (0:00:00.243)       0:00:57.589 ******** \n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'})\u001b[0m\n\nTASK [Print 'Create block VGs'] ************************************************\nSaturday 21 March 2026  00:43:13 +0000 (0:00:01.829)       0:00:59.419 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create block LVs] ********************************************************\nSaturday 21 March 2026  00:43:13 +0000 (0:00:00.166)       0:00:59.586 ******** \n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'})\u001b[0m\n\nTASK [Print 'Create block LVs'] ************************************************\nSaturday 21 March 2026  00:43:15 +0000 (0:00:01.303)       0:01:00.889 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create DB VGs] ***********************************************************\nSaturday 21 March 2026  00:43:15 +0000 (0:00:00.166)       0:01:01.055 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create DB VGs'] ***************************************************\nSaturday 21 March 2026  00:43:15 +0000 (0:00:00.129)       0:01:01.185 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create WAL VGs] **********************************************************\nSaturday 21 March 2026  00:43:15 +0000 (0:00:00.152)       0:01:01.338 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create WAL VGs'] **************************************************\nSaturday 21 March 2026  00:43:15 +0000 (0:00:00.135)       0:01:01.473 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create DB+WAL VGs] *******************************************************\nSaturday 21 March 2026  00:43:15 +0000 (0:00:00.141)       0:01:01.614 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create DB+WAL VGs'] ***********************************************\nSaturday 21 March 2026  00:43:15 +0000 (0:00:00.131)       0:01:01.746 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Prepare variables for OSD count check] ***********************************\nSaturday 21 March 2026  00:43:16 +0000 (0:00:00.157)       0:01:01.904 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************\nSaturday 21 March 2026  00:43:16 +0000 (0:00:00.134)       0:01:02.038 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************\nSaturday 21 March 2026  00:43:16 +0000 (0:00:00.337)       0:01:02.376 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************\nSaturday 21 March 2026  00:43:16 +0000 (0:00:00.161)       0:01:02.537 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************\nSaturday 21 March 2026  00:43:16 +0000 (0:00:00.153)       0:01:02.690 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************\nSaturday 21 March 2026  00:43:17 +0000 (0:00:00.139)       0:01:02.830 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************\nSaturday 21 March 2026  00:43:17 +0000 (0:00:00.138)       0:01:02.968 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print number of OSDs wanted per DB VG] ***********************************\nSaturday 21 March 2026  00:43:17 +0000 (0:00:00.126)       0:01:03.094 ******** \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per WAL VG] **********************************\nSaturday 21 March 2026  00:43:17 +0000 (0:00:00.139)       0:01:03.234 ******** \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per DB+WAL VG] *******************************\nSaturday 21 March 2026  00:43:17 +0000 (0:00:00.139)       0:01:03.374 ******** \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Gather DB VGs with total and available size in bytes] ********************\nSaturday 21 March 2026  00:43:17 +0000 (0:00:00.142)       0:01:03.516 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Gather WAL VGs with total and available size in bytes] *******************\nSaturday 21 March 2026  00:43:18 +0000 (0:00:00.550)       0:01:04.067 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Gather DB+WAL VGs with total and available size in bytes] ****************\nSaturday 21 March 2026  00:43:18 +0000 (0:00:00.574)       0:01:04.642 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************\nSaturday 21 March 2026  00:43:19 +0000 (0:00:00.509)       0:01:05.151 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Calculate VG sizes (without buffer)] *************************************\nSaturday 21 March 2026  00:43:19 +0000 (0:00:00.337)       0:01:05.489 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate VG sizes (with buffer)] ****************************************\nSaturday 21 March 2026  00:43:19 +0000 (0:00:00.103)       0:01:05.593 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print LVM VGs report data] ***********************************************\nSaturday 21 March 2026  00:43:19 +0000 (0:00:00.106)       0:01:05.699 ******** \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"vgs_report\": {\u001b[0m\n\u001b[0;32m        \"vg\": []\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print LVM VG sizes] ******************************************************\nSaturday 21 March 2026  00:43:20 +0000 (0:00:00.139)       0:01:05.838 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_db_devices] ************************\nSaturday 21 March 2026  00:43:20 +0000 (0:00:00.139)       0:01:05.978 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_db_devices] ****************************\nSaturday 21 March 2026  00:43:20 +0000 (0:00:00.133)       0:01:06.112 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if size of DB LVs on ceph_db_devices > available] *******************\nSaturday 21 March 2026  00:43:20 +0000 (0:00:00.132)       0:01:06.244 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_wal_devices] ***********************\nSaturday 21 March 2026  00:43:20 +0000 (0:00:00.137)       0:01:06.381 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_wal_devices] ***************************\nSaturday 21 March 2026  00:43:20 +0000 (0:00:00.134)       0:01:06.516 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************\nSaturday 21 March 2026  00:43:20 +0000 (0:00:00.136)       0:01:06.652 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************\nSaturday 21 March 2026  00:43:20 +0000 (0:00:00.110)       0:01:06.762 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************\nSaturday 21 March 2026  00:43:21 +0000 (0:00:00.129)       0:01:06.892 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************\nSaturday 21 March 2026  00:43:21 +0000 (0:00:00.338)       0:01:07.230 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************\nSaturday 21 March 2026  00:43:21 +0000 (0:00:00.139)       0:01:07.370 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********\nSaturday 21 March 2026  00:43:21 +0000 (0:00:00.138)       0:01:07.508 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************\nSaturday 21 March 2026  00:43:21 +0000 (0:00:00.137)       0:01:07.645 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************\nSaturday 21 March 2026  00:43:21 +0000 (0:00:00.145)       0:01:07.791 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_devices] ***************************************\nSaturday 21 March 2026  00:43:22 +0000 (0:00:00.139)       0:01:07.931 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_devices'] *******************************\nSaturday 21 March 2026  00:43:22 +0000 (0:00:00.151)       0:01:08.082 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create WAL LVs for ceph_wal_devices] *************************************\nSaturday 21 March 2026  00:43:22 +0000 (0:00:00.153)       0:01:08.235 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************\nSaturday 21 March 2026  00:43:22 +0000 (0:00:00.160)       0:01:08.396 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create WAL LVs for ceph_db_wal_devices] **********************************\nSaturday 21 March 2026  00:43:22 +0000 (0:00:00.146)       0:01:08.543 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************\nSaturday 21 March 2026  00:43:22 +0000 (0:00:00.158)       0:01:08.701 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_wal_devices] ***********************************\nSaturday 21 March 2026  00:43:23 +0000 (0:00:00.153)       0:01:08.854 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************\nSaturday 21 March 2026  00:43:23 +0000 (0:00:00.359)       0:01:09.214 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Get list of Ceph LVs with associated VGs] ********************************\nSaturday 21 March 2026  00:43:23 +0000 (0:00:00.160)       0:01:09.374 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get list of Ceph PVs with associated VGs] ********************************\nSaturday 21 March 2026  00:43:24 +0000 (0:00:00.530)       0:01:09.904 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************\nSaturday 21 March 2026  00:43:24 +0000 (0:00:00.569)       0:01:10.473 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Create list of VG/LV names] **********************************************\nSaturday 21 March 2026  00:43:24 +0000 (0:00:00.138)       0:01:10.612 ******** \n\u001b[0;32mok: [testbed-node-5] => (item={'lv_name': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'vg_name': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'lv_name': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'vg_name': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'})\u001b[0m\n\nTASK [Fail if block LV defined in lvm_volumes is missing] **********************\nSaturday 21 March 2026  00:43:24 +0000 (0:00:00.170)       0:01:10.782 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if DB LV defined in lvm_volumes is missing] *************************\nSaturday 21 March 2026  00:43:25 +0000 (0:00:00.155)       0:01:10.937 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if WAL LV defined in lvm_volumes is missing] ************************\nSaturday 21 March 2026  00:43:25 +0000 (0:00:00.157)       0:01:11.095 ******** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print LVM report data] ***************************************************\nSaturday 21 March 2026  00:43:25 +0000 (0:00:00.172)       0:01:11.267 ******** \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"lvm_report\": {\u001b[0m\n\u001b[0;32m        \"lv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-ede169e1-d974-55ed-a938-c303cb89cb5b\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ],\u001b[0m\n\u001b[0;32m        \"pv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-ede169e1-d974-55ed-a938-c303cb89cb5b\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ]\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:43:25 +0000 (0:00:00.145)       0:01:11.413 ******** \n=============================================================================== \nCreate block VGs -------------------------------------------------------- 5.70s\nCreate block LVs -------------------------------------------------------- 3.95s\nGather DB VGs with total and available size in bytes -------------------- 1.99s\nGather WAL VGs with total and available size in bytes ------------------- 1.66s\nGather DB+WAL VGs with total and available size in bytes ---------------- 1.65s\nGet list of Ceph PVs with associated VGs -------------------------------- 1.60s\nGet list of Ceph LVs with associated VGs -------------------------------- 1.54s\nAdd known partitions to the list of available block devices ------------- 1.38s\nAdd known partitions to the list of available block devices ------------- 1.37s\nAdd known links to the list of available block devices ------------------ 1.15s\nPrint LVM report data --------------------------------------------------- 0.85s\nAdd known links to the list of available block devices ------------------ 0.82s\nPrint number of OSDs wanted per DB VG ----------------------------------- 0.76s\nAdd known partitions to the list of available block devices ------------- 0.74s\nGet extra vars for Ceph configuration ----------------------------------- 0.72s\nPrint 'Create DB LVs for ceph_db_devices' ------------------------------- 0.70s\nPrint 'Create WAL LVs for ceph_wal_devices' ----------------------------- 0.69s\nCount OSDs put on ceph_db_devices defined in lvm_volumes ---------------- 0.69s\nFail if DB LV defined in lvm_volumes is missing ------------------------- 0.67s\nGet initial list of available block devices ----------------------------- 0.67s\n\u001b[32m2026-03-21 00:43:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of facts.\u001b[0m\n\u001b[32m2026-03-21 00:43:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9c028823-76e0-4a79-8ec2-6e01560eaf6e (facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 00:43:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 9c028823-76e0-4a79-8ec2-6e01560eaf6e (facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role facts] ********************************************************\n\nTASK [osism.commons.facts : Create custom facts directory] *********************\nSaturday 21 March 2026  00:43:40 +0000 (0:00:00.328)       0:00:00.328 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.facts : Copy fact files] ***********************************\nSaturday 21 March 2026  00:43:42 +0000 (0:00:01.265)       0:00:01.594 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nSaturday 21 March 2026  00:43:43 +0000 (0:00:01.063)       0:00:02.657 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts if using --limit] *****************************\n\nTASK [Gather facts for all hosts] **********************************************\nSaturday 21 March 2026  00:43:48 +0000 (0:00:04.876)       0:00:07.534 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:43:48 +0000 (0:00:00.500)       0:00:08.035 ******** \n=============================================================================== \nGathers facts about hosts ----------------------------------------------- 4.88s\nosism.commons.facts : Create custom facts directory --------------------- 1.27s\nosism.commons.facts : Copy fact files ----------------------------------- 1.06s\nGather facts for all hosts ---------------------------------------------- 0.50s\n\u001b[32m2026-03-21 00:44:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of frr.\u001b[0m\n\u001b[32m2026-03-21 00:44:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dc4b529e-724d-4a2f-88cc-989a0ee9615c (frr) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 00:44:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task dc4b529e-724d-4a2f-88cc-989a0ee9615c (frr) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role frr] **********************************************************\n\nTASK [osism.services.frr : Include distribution specific install tasks] ********\nSaturday 21 March 2026  00:44:02 +0000 (0:00:00.227)       0:00:00.227 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/frr/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.services.frr : Pin frr package version] ****************************\nSaturday 21 March 2026  00:44:03 +0000 (0:00:00.213)       0:00:00.441 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Install frr package] ********************************\nSaturday 21 March 2026  00:44:04 +0000 (0:00:01.395)       0:00:01.837 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Copy file: /etc/frr/vtysh.conf] *********************\nSaturday 21 March 2026  00:44:13 +0000 (0:00:08.534)       0:00:10.372 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Copy file: /etc/frr/daemons] ************************\nSaturday 21 March 2026  00:44:13 +0000 (0:00:00.944)       0:00:11.316 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Set _frr_uplinks fact] ******************************\nSaturday 21 March 2026  00:44:14 +0000 (0:00:00.829)       0:00:12.146 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Write frr_config_template to temporary file] ********\nSaturday 21 March 2026  00:44:15 +0000 (0:00:01.108)       0:00:13.255 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Render frr.conf from frr_config_template variable] ***\nSaturday 21 March 2026  00:44:16 +0000 (0:00:00.149)       0:00:13.404 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Remove temporary frr_config_template file] **********\nSaturday 21 March 2026  00:44:16 +0000 (0:00:00.255)       0:00:13.659 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Check for frr.conf file in the configuration repository] ***\nSaturday 21 March 2026  00:44:16 +0000 (0:00:00.148)       0:00:13.808 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Copy frr.conf file from the configuration repository] ***\nSaturday 21 March 2026  00:44:16 +0000 (0:00:00.132)       0:00:13.941 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Copy default frr.conf file of type k3s_cilium] ******\nSaturday 21 March 2026  00:44:16 +0000 (0:00:00.143)       0:00:14.084 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Set sysctl parameters] ******************************\nSaturday 21 March 2026  00:44:17 +0000 (0:00:00.893)       0:00:14.978 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.ip_forward', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.send_redirects', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.accept_redirects', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.fib_multipath_hash_policy', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.default.ignore_routes_with_linkdown', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.rp_filter', 'value': 2})\u001b[0m\n\nTASK [osism.services.frr : Manage frr service] *********************************\nSaturday 21 March 2026  00:44:19 +0000 (0:00:02.033)       0:00:17.012 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.frr : Restart frr service] *********************\nSaturday 21 March 2026  00:44:20 +0000 (0:00:01.100)       0:00:18.112 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:44:22 +0000 (0:00:01.315)       0:00:19.428 ******** \n=============================================================================== \nosism.services.frr : Install frr package -------------------------------- 8.53s\nosism.services.frr : Set sysctl parameters ------------------------------ 2.03s\nosism.services.frr : Pin frr package version ---------------------------- 1.40s\nosism.services.frr : Restart frr service -------------------------------- 1.32s\nosism.services.frr : Set _frr_uplinks fact ------------------------------ 1.11s\nosism.services.frr : Manage frr service --------------------------------- 1.10s\nosism.services.frr : Copy file: /etc/frr/vtysh.conf --------------------- 0.94s\nosism.services.frr : Copy default frr.conf file of type k3s_cilium ------ 0.89s\nosism.services.frr : Copy file: /etc/frr/daemons ------------------------ 0.83s\nosism.services.frr : Render frr.conf from frr_config_template variable --- 0.26s\nosism.services.frr : Include distribution specific install tasks -------- 0.21s\nosism.services.frr : Write frr_config_template to temporary file -------- 0.15s\nosism.services.frr : Remove temporary frr_config_template file ---------- 0.15s\nosism.services.frr : Copy frr.conf file from the configuration repository --- 0.14s\nosism.services.frr : Check for frr.conf file in the configuration repository --- 0.13s\n\n--> DEPLOY IN A NUTSHELL -- START -- Sat Mar 21 00:44:22 UTC 2026\n\n\u001b[32m2026-03-21 00:44:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCollection nutshell is prepared for execution\u001b[0m\n\u001b[32m2026-03-21 00:44:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - dotfiles\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - homer\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - netdata\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - openstackclient\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - phpmyadmin\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - common\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- loadbalancer\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- opensearch\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- mariadb-ng\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [3] ---- horizon\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [3] ---- keystone\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- neutron\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ wait-for-nova\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [6] ------- octavia\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- barbican\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- designate\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- ironic\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- placement\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- magnum\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- openvswitch\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- ovn\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- memcached\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- redis\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- rabbitmq-ng\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - kubernetes\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- kubeconfig\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- copy-kubeconfig\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - ceph\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- ceph-pools\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- copy-ceph-keys\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [3] ---- cephclient\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- ceph-bootstrap-dashboard\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- wait-for-keystone\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ kolla-ceph-rgw\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ glance\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ cinder\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ nova\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- prometheus\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ grafana\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAll tasks of the collection nutshell are prepared for execution\u001b[0m\n\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTasks are running in the background\u001b[0m\n\u001b[32m2026-03-21 00:44:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mNo task IDs specified, wait for all currently running tasks\u001b[0m\n\u001b[32m2026-03-21 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask be72019d-4861-4774-8ce4-d34fe16c1c55 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask be72019d-4861-4774-8ce4-d34fe16c1c55 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask be72019d-4861-4774-8ce4-d34fe16c1c55 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:44:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask be72019d-4861-4774-8ce4-d34fe16c1c55 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:44:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask be72019d-4861-4774-8ce4-d34fe16c1c55 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:44:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask be72019d-4861-4774-8ce4-d34fe16c1c55 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:44:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask be72019d-4861-4774-8ce4-d34fe16c1c55 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\nPLAY [Apply role geerlingguy.dotfiles] *****************************************\n\nTASK [geerlingguy.dotfiles : Ensure dotfiles repository is cloned locally.] ****\nSaturday 21 March 2026  00:44:44 +0000 (0:00:00.673)       0:00:00.673 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [geerlingguy.dotfiles : Ensure all configured dotfiles are links.] ********\nSaturday 21 March 2026  00:44:49 +0000 (0:00:05.308)       0:00:05.982 ******** \n\u001b[0;32mok: [testbed-manager] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=.tmux.conf)\u001b[0m\n\nTASK [geerlingguy.dotfiles : Remove existing dotfiles file if a replacement is being linked.] ***\nSaturday 21 March 2026  00:44:52 +0000 (0:00:02.156)       0:00:08.138 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-21 00:44:50.481228', 'end': '2026-03-21 00:44:51.488814', 'delta': '0:00:01.007586', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-21 00:44:50.685102', 'end': '2026-03-21 00:44:50.691209', 'delta': '0:00:00.006107', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-21 00:44:50.266670', 'end': '2026-03-21 00:44:50.270757', 'delta': '0:00:00.004087', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-21 00:44:50.972699', 'end': '2026-03-21 00:44:50.979364', 'delta': '0:00:00.006665', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-21 00:44:51.476165', 'end': '2026-03-21 00:44:51.485035', 'delta': '0:00:00.008870', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-21 00:44:51.684500', 'end': '2026-03-21 00:44:51.690397', 'delta': '0:00:00.005897', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-21 00:44:50.964795', 'end': '2026-03-21 00:44:50.970733', 'delta': '0:00:00.005938', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\nTASK [geerlingguy.dotfiles : Ensure parent folders of link dotfiles exist.] ****\nSaturday 21 March 2026  00:44:53 +0000 (0:00:01.376)       0:00:09.515 ******** \n\u001b[0;32mok: [testbed-manager] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=.tmux.conf)\u001b[0m\n\nTASK [geerlingguy.dotfiles : Link dotfiles into home folder.] ******************\nSaturday 21 March 2026  00:44:54 +0000 (0:00:01.332)       0:00:10.847 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=.tmux.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=.tmux.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=.tmux.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=.tmux.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=.tmux.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=.tmux.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=.tmux.conf)\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:44:57 +0000 (0:00:02.576)       0:00:13.424 ******** \n=============================================================================== \ngeerlingguy.dotfiles : Ensure dotfiles repository is cloned locally. ---- 5.31s\ngeerlingguy.dotfiles : Link dotfiles into home folder. ------------------ 2.58s\ngeerlingguy.dotfiles : Ensure all configured dotfiles are links. -------- 2.16s\ngeerlingguy.dotfiles : Remove existing dotfiles file if a replacement is being linked. --- 1.38s\ngeerlingguy.dotfiles : Ensure parent folders of link dotfiles exist. ---- 1.33s\n\u001b[32m2026-03-21 00:44:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask be72019d-4861-4774-8ce4-d34fe16c1c55 is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-21 00:44:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:44:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:45:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:45:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:45:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:45:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:45:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:45:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:45:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:45:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:45:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:45:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-21 00:45:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:45:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-21 00:45:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:45:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:45:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:45:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:45:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:45:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:45:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:45:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:45:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:45:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:46:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:46:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:46:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:46:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:46:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state SUCCESS\u001b[0m\n\n\nPLAY [Apply role homer] ********************************************************\n\nTASK [osism.services.homer : Inform about new parameter homer_url_opensearch_dashboards] ***\nSaturday 21 March 2026  00:44:43 +0000 (0:00:00.884)       0:00:00.884 ******** \n\u001b[0;32mok: [testbed-manager] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"The support for the homer_url_kibana has been removed. Please use the homer_url_opensearch_dashboards parameter.\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [osism.services.homer : Create traefik external network] ******************\nSaturday 21 March 2026  00:44:44 +0000 (0:00:00.369)       0:00:01.253 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.homer : Create required directories] **********************\nSaturday 21 March 2026  00:44:47 +0000 (0:00:03.616)       0:00:04.870 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/homer/configuration)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/homer)\u001b[0m\n\nTASK [osism.services.homer : Copy config.yml configuration file] ***************\nSaturday 21 March 2026  00:44:49 +0000 (0:00:01.718)       0:00:06.588 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.homer : Copy docker-compose.yml file] *********************\nSaturday 21 March 2026  00:44:51 +0000 (0:00:02.319)       0:00:08.907 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.homer : Manage homer service] *****************************\nSaturday 21 March 2026  00:44:53 +0000 (0:00:01.883)       0:00:10.792 ******** \n\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage homer service (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.homer : Restart homer service] *****************\nSaturday 21 March 2026  00:45:21 +0000 (0:00:28.166)       0:00:38.958 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:45:27 +0000 (0:00:05.262)       0:00:44.221 ******** \n=============================================================================== \nosism.services.homer : Manage homer service ---------------------------- 28.17s\nosism.services.homer : Restart homer service ---------------------------- 5.26s\nosism.services.homer : Create traefik external network ------------------ 3.62s\nosism.services.homer : Copy config.yml configuration file --------------- 2.32s\nosism.services.homer : Copy docker-compose.yml file --------------------- 1.88s\nosism.services.homer : Create required directories ---------------------- 1.72s\nosism.services.homer : Inform about new parameter homer_url_opensearch_dashboards --- 0.37s\n\n\nPLAY [Apply role openstackclient] **********************************************\n\nTASK [osism.services.openstackclient : Include tasks] **************************\nSaturday 21 March 2026  00:44:44 +0000 (0:00:01.129)       0:00:01.129 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/openstackclient/tasks/container-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.services.openstackclient : Create required directories] ************\nSaturday 21 March 2026  00:44:45 +0000 (0:00:00.769)       0:00:01.899 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/configuration/environments/openstack)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/openstackclient/data)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/openstackclient)\u001b[0m\n\nTASK [osism.services.openstackclient : Copy docker-compose.yml file] ***********\nSaturday 21 March 2026  00:44:48 +0000 (0:00:02.972)       0:00:04.872 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.openstackclient : Manage openstackclient service] *********\nSaturday 21 March 2026  00:44:50 +0000 (0:00:02.103)       0:00:06.976 ******** \n\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage openstackclient service (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.openstackclient : Copy openstack wrapper script] **********\nSaturday 21 March 2026  00:45:25 +0000 (0:00:35.346)       0:00:42.322 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.openstackclient : Remove ospurge wrapper script] **********\nSaturday 21 March 2026  00:45:26 +0000 (0:00:00.995)       0:00:43.318 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.openstackclient : Restart openstackclient service] ***\nSaturday 21 March 2026  00:45:27 +0000 (0:00:01.216)       0:00:44.535 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.openstackclient : Ensure that all containers are up] ***\nSaturday 21 March 2026  00:45:29 +0000 (0:00:01.654)       0:00:46.189 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.openstackclient : Wait for an healthy service] ***\nSaturday 21 March 2026  00:45:30 +0000 (0:00:01.476)       0:00:47.665 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.openstackclient : Copy bash completion script] ***\nSaturday 21 March 2026  00:45:31 +0000 (0:00:00.887)       0:00:48.553 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:45:32 +0000 (0:00:00.512)       0:00:49.066 ******** \n=============================================================================== \nosism.services.openstackclient : Manage openstackclient service -------- 35.35s\nosism.services.openstackclient : Create required directories ------------ 2.97s\nosism.services.openstackclient : Copy docker-compose.yml file ----------- 2.10s\nosism.services.openstackclient : Restart openstackclient service -------- 1.65s\nosism.services.openstackclient : Ensure that all containers are up ------ 1.48s\nosism.services.openstackclient : Remove ospurge wrapper script ---------- 1.22s\nosism.services.openstackclient : Copy openstack wrapper script ---------- 1.00s\nosism.services.openstackclient : Wait for an healthy service ------------ 0.89s\nosism.services.openstackclient : Include tasks -------------------------- 0.77s\nosism.services.openstackclient : Copy bash completion script ------------ 0.51s\n\n\nPLAY [Apply role phpmyadmin] ***************************************************\n\nTASK [osism.services.phpmyadmin : Create traefik external network] *************\nSaturday 21 March 2026  00:45:05 +0000 (0:00:00.405)       0:00:00.405 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.phpmyadmin : Create required directories] *****************\nSaturday 21 March 2026  00:45:07 +0000 (0:00:01.492)       0:00:01.897 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/phpmyadmin)\u001b[0m\n\nTASK [osism.services.phpmyadmin : Copy docker-compose.yml file] ****************\nSaturday 21 March 2026  00:45:08 +0000 (0:00:00.765)       0:00:02.663 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.phpmyadmin : Manage phpmyadmin service] *******************\nSaturday 21 March 2026  00:45:09 +0000 (0:00:01.645)       0:00:04.309 ******** \n\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage phpmyadmin service (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.phpmyadmin : Restart phpmyadmin service] *******\nSaturday 21 March 2026  00:46:09 +0000 (0:00:59.909)       0:01:04.218 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:46:12 +0000 (0:00:02.982)       0:01:07.201 ******** \n=============================================================================== \nosism.services.phpmyadmin : Manage phpmyadmin service ------------------ 59.91s\nosism.services.phpmyadmin : Restart phpmyadmin service ------------------ 2.98s\nosism.services.phpmyadmin : Copy docker-compose.yml file ---------------- 1.65s\nosism.services.phpmyadmin : Create traefik external network ------------- 1.49s\nosism.services.phpmyadmin : Create required directories ----------------- 0.77s\n\u001b[32m2026-03-21 00:46:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:46:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  00:44:44 +0000 (0:00:00.844)       0:00:00.844 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=enable_netdata_True)\u001b[0m\n\nPLAY [Apply role netdata] ******************************************************\n\nTASK [osism.services.netdata : Include distribution specific install tasks] ****\nSaturday 21 March 2026  00:44:46 +0000 (0:00:01.737)       0:00:02.581 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.netdata : Remove old architecture-dependent repository] ***\nSaturday 21 March 2026  00:44:48 +0000 (0:00:02.214)       0:00:04.796 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Install apt-transport-https package] ************\nSaturday 21 March 2026  00:44:50 +0000 (0:00:02.606)       0:00:07.402 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.services.netdata : Add repository gpg key] *************************\nSaturday 21 March 2026  00:44:55 +0000 (0:00:04.300)       0:00:11.702 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Add repository] *********************************\nSaturday 21 March 2026  00:44:57 +0000 (0:00:02.212)       0:00:13.915 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.netdata : Install package netdata] ************************\nSaturday 21 March 2026  00:45:08 +0000 (0:00:10.636)       0:00:24.552 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.netdata : Include config tasks] ***************************\nSaturday 21 March 2026  00:45:44 +0000 (0:00:36.554)       0:01:01.107 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/config.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.netdata : Copy configuration files] ***********************\nSaturday 21 March 2026  00:45:46 +0000 (0:00:02.024)       0:01:03.131 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=stream.conf)\u001b[0m\n\nTASK [osism.services.netdata : Retrieve /etc/netdata/.opt-out-from-anonymous-statistics status] ***\nSaturday 21 March 2026  00:45:52 +0000 (0:00:06.145)       0:01:09.277 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Opt out from anonymous statistics] **************\nSaturday 21 March 2026  00:45:54 +0000 (0:00:01.470)       0:01:10.747 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Add netdata user to docker group] ***************\nSaturday 21 March 2026  00:45:56 +0000 (0:00:01.899)       0:01:12.647 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Manage service netdata] *************************\nSaturday 21 March 2026  00:45:57 +0000 (0:00:01.641)       0:01:14.289 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Include host type specific tasks] ***************\nSaturday 21 March 2026  00:45:59 +0000 (0:00:01.904)       0:01:16.193 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/server.yml for testbed-manager\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/client.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.netdata : Set sysctl vm.max_map_count parameter] **********\nSaturday 21 March 2026  00:46:01 +0000 (0:00:01.416)       0:01:17.610 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.netdata : Restart service netdata] *************\nSaturday 21 March 2026  00:46:03 +0000 (0:00:02.636)       0:01:20.246 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:46:15 +0000 (0:00:11.497)       0:01:31.743 ******** \n=============================================================================== \nosism.services.netdata : Install package netdata ----------------------- 36.55s\nosism.services.netdata : Restart service netdata ----------------------- 11.50s\nosism.services.netdata : Add repository -------------------------------- 10.64s\nosism.services.netdata : Copy configuration files ----------------------- 6.15s\nosism.services.netdata : Install apt-transport-https package ------------ 4.30s\nosism.services.netdata : Set sysctl vm.max_map_count parameter ---------- 2.64s\nosism.services.netdata : Remove old architecture-dependent repository --- 2.61s\nosism.services.netdata : Include distribution specific install tasks ---- 2.21s\nosism.services.netdata : Add repository gpg key ------------------------- 2.21s\nosism.services.netdata : Include config tasks --------------------------- 2.02s\nosism.services.netdata : Manage service netdata ------------------------- 1.90s\nosism.services.netdata : Opt out from anonymous statistics -------------- 1.90s\nGroup hosts based on enabled services ----------------------------------- 1.74s\nosism.services.netdata : Add netdata user to docker group --------------- 1.64s\nosism.services.netdata : Retrieve /etc/netdata/.opt-out-from-anonymous-statistics status --- 1.47s\nosism.services.netdata : Include host type specific tasks --------------- 1.42s\n\u001b[32m2026-03-21 00:46:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-21 00:46:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:46:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:46:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:46:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:46:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:46:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:46:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:46:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:46:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:46:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:46:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:46:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:46:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:46:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:46:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:46:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:47:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:47:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:47:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:47:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:47:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c076440c-db82-4d81-8e70-68ed60306a74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state SUCCESS\u001b[0m\n\n\nPLAY [Apply role common] *******************************************************\n\nTASK [common : include_tasks] **************************************************\nSaturday 21 March 2026  00:44:37 +0000 (0:00:00.282)       0:00:00.282 ******** \n\u001b[0;36mincluded: /ansible/roles/common/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [common : Ensuring config directories exist] ******************************\nSaturday 21 March 2026  00:44:38 +0000 (0:00:01.310)       0:00:01.593 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\nTASK [common : include_tasks] **************************************************\nSaturday 21 March 2026  00:44:42 +0000 (0:00:03.850)       0:00:05.443 ******** \n\u001b[0;36mincluded: /ansible/roles/common/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [service-cert-copy : common | Copying over extra CA certificates] *********\nSaturday 21 March 2026  00:44:44 +0000 (0:00:01.416)       0:00:06.860 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-cert-copy : common | Copying over backend internal TLS certificate] ***\nSaturday 21 March 2026  00:44:48 +0000 (0:00:04.699)       0:00:11.559 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [service-cert-copy : common | Copying over backend internal TLS key] ******\nSaturday 21 March 2026  00:44:50 +0000 (0:00:02.052)       0:00:13.612 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Copying over /run subdirectories conf] **************************\nSaturday 21 March 2026  00:44:53 +0000 (0:00:02.440)       0:00:16.053 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Restart systemd-tmpfiles] ***************************************\nSaturday 21 March 2026  00:44:54 +0000 (0:00:01.227)       0:00:17.280 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Copying over config.json files for services] ********************\nSaturday 21 March 2026  00:44:55 +0000 (0:00:01.368)       0:00:18.648 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [common : Find custom fluentd input config files] *************************\nSaturday 21 March 2026  00:45:04 +0000 (0:00:08.764)       0:00:27.413 ******** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Find custom fluentd filter config files] ************************\nSaturday 21 March 2026  00:45:06 +0000 (0:00:01.451)       0:00:28.864 ******** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Find custom fluentd format config files] ************************\nSaturday 21 March 2026  00:45:07 +0000 (0:00:00.904)       0:00:29.768 ******** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Find custom fluentd output config files] ************************\nSaturday 21 March 2026  00:45:07 +0000 (0:00:00.869)       0:00:30.637 ******** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Copying over fluentd.conf] **************************************\nSaturday 21 March 2026  00:45:08 +0000 (0:00:00.942)       0:00:31.580 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [common : Copying over cron logrotate config file] ************************\nSaturday 21 March 2026  00:45:13 +0000 (0:00:04.842)       0:00:36.422 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\nTASK [common : Ensure RabbitMQ Erlang cookie exists] ***************************\nSaturday 21 March 2026  00:45:17 +0000 (0:00:03.422)       0:00:39.845 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [common : Ensuring config directories have correct owner and permission] ***\nSaturday 21 March 2026  00:45:19 +0000 (0:00:02.339)       0:00:42.184 ******** \n\u001b[0;32mok: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [common : Copy rabbitmq-env.conf to kolla toolbox] ************************\nSaturday 21 March 2026  00:45:21 +0000 (0:00:02.295)       0:00:44.480 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\nTASK [common : Copy rabbitmq erl_inetrc to kolla toolbox] **********************\nSaturday 21 March 2026  00:45:26 +0000 (0:00:04.327)       0:00:48.808 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\nTASK [common : Check common containers] ****************************************\nSaturday 21 March 2026  00:45:29 +0000 (0:00:03.615)       0:00:52.424 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [common : Creating log volume] ********************************************\nSaturday 21 March 2026  00:45:33 +0000 (0:00:03.671)       0:00:56.096 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [common : Link kolla_logs volume to /var/log/kolla] ***********************\nSaturday 21 March 2026  00:45:35 +0000 (0:00:01.766)       0:00:57.862 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [common : Flush handlers] *************************************************\nSaturday 21 March 2026  00:45:36 +0000 (0:00:01.511)       0:00:59.374 ******** \n\nTASK [common : Flush handlers] *************************************************\nSaturday 21 March 2026  00:45:36 +0000 (0:00:00.073)       0:00:59.447 ******** \n\nTASK [common : Flush handlers] *************************************************\nSaturday 21 March 2026  00:45:36 +0000 (0:00:00.067)       0:00:59.514 ******** \n\nTASK [common : Flush handlers] *************************************************\nSaturday 21 March 2026  00:45:36 +0000 (0:00:00.075)       0:00:59.590 ******** \n\nTASK [common : Flush handlers] *************************************************\nSaturday 21 March 2026  00:45:36 +0000 (0:00:00.083)       0:00:59.673 ******** \n\nTASK [common : Flush handlers] *************************************************\nSaturday 21 March 2026  00:45:37 +0000 (0:00:00.068)       0:00:59.742 ******** \n\nTASK [common : Flush handlers] *************************************************\nSaturday 21 March 2026  00:45:37 +0000 (0:00:00.076)       0:00:59.818 ******** \n\nRUNNING HANDLER [common : Restart fluentd container] ***************************\nSaturday 21 March 2026  00:45:37 +0000 (0:00:00.092)       0:00:59.910 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [common : Restart kolla-toolbox container] *********************\nSaturday 21 March 2026  00:46:12 +0000 (0:00:34.955)       0:01:34.866 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [common : Initializing toolbox container using normal user] ****\nSaturday 21 March 2026  00:47:03 +0000 (0:00:51.027)       0:02:25.894 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [common : Restart cron container] ******************************\nSaturday 21 March 2026  00:47:04 +0000 (0:00:01.779)       0:02:27.673 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:47:14 +0000 (0:00:09.344)       0:02:37.018 ******** \n=============================================================================== \ncommon : Restart kolla-toolbox container ------------------------------- 51.03s\ncommon : Restart fluentd container ------------------------------------- 34.96s\ncommon : Restart cron container ----------------------------------------- 9.34s\ncommon : Copying over config.json files for services -------------------- 8.76s\ncommon : Copying over fluentd.conf -------------------------------------- 4.84s\nservice-cert-copy : common | Copying over extra CA certificates --------- 4.70s\ncommon : Copy rabbitmq-env.conf to kolla toolbox ------------------------ 4.33s\ncommon : Ensuring config directories exist ------------------------------ 3.85s\ncommon : Check common containers ---------------------------------------- 3.67s\ncommon : Copy rabbitmq erl_inetrc to kolla toolbox ---------------------- 3.62s\ncommon : Copying over cron logrotate config file ------------------------ 3.42s\nservice-cert-copy : common | Copying over backend internal TLS key ------ 2.44s\ncommon : Ensure RabbitMQ Erlang cookie exists --------------------------- 2.34s\ncommon : Ensuring config directories have correct owner and permission --- 2.30s\nservice-cert-copy : common | Copying over backend internal TLS certificate --- 2.05s\ncommon : Initializing toolbox container using normal user --------------- 1.78s\ncommon : Creating log volume -------------------------------------------- 1.77s\ncommon : Link kolla_logs volume to /var/log/kolla ----------------------- 1.51s\ncommon : Find custom fluentd input config files ------------------------- 1.45s\ncommon : include_tasks -------------------------------------------------- 1.42s\n\u001b[32m2026-03-21 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0d652c86-9d68-4200-b5b9-a2d094b18b9c is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c076440c-db82-4d81-8e70-68ed60306a74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0d652c86-9d68-4200-b5b9-a2d094b18b9c is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c076440c-db82-4d81-8e70-68ed60306a74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0d652c86-9d68-4200-b5b9-a2d094b18b9c is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c076440c-db82-4d81-8e70-68ed60306a74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0d652c86-9d68-4200-b5b9-a2d094b18b9c is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c076440c-db82-4d81-8e70-68ed60306a74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0d652c86-9d68-4200-b5b9-a2d094b18b9c is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:47:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c076440c-db82-4d81-8e70-68ed60306a74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0d652c86-9d68-4200-b5b9-a2d094b18b9c is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:47:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c076440c-db82-4d81-8e70-68ed60306a74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0d652c86-9d68-4200-b5b9-a2d094b18b9c is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-21 00:47:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:47:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c076440c-db82-4d81-8e70-68ed60306a74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c076440c-db82-4d81-8e70-68ed60306a74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c076440c-db82-4d81-8e70-68ed60306a74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c076440c-db82-4d81-8e70-68ed60306a74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  00:47:19 +0000 (0:00:00.552)       0:00:00.552 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  00:47:20 +0000 (0:00:00.643)       0:00:01.196 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_memcached_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_memcached_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_memcached_True)\u001b[0m\n\nPLAY [Apply role memcached] ****************************************************\n\nTASK [memcached : include_tasks] ***********************************************\nSaturday 21 March 2026  00:47:20 +0000 (0:00:00.477)       0:00:01.673 ******** \n\u001b[0;36mincluded: /ansible/roles/memcached/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [memcached : Ensuring config directories exist] ***************************\nSaturday 21 March 2026  00:47:21 +0000 (0:00:00.531)       0:00:02.205 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=memcached)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=memcached)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=memcached)\u001b[0m\n\nTASK [memcached : Copying over config.json files for services] *****************\nSaturday 21 March 2026  00:47:22 +0000 (0:00:01.730)       0:00:03.935 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item=memcached)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=memcached)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=memcached)\u001b[0m\n\nTASK [memcached : Check memcached container] ***********************************\nSaturday 21 March 2026  00:47:25 +0000 (0:00:02.216)       0:00:06.151 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [memcached : Restart memcached container] **********************\nSaturday 21 March 2026  00:47:27 +0000 (0:00:02.571)       0:00:08.722 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:47:34 +0000 (0:00:06.581)       0:00:15.304 ******** \n=============================================================================== \nmemcached : Restart memcached container --------------------------------- 6.58s\nmemcached : Check memcached container ----------------------------------- 2.57s\nmemcached : Copying over config.json files for services ----------------- 2.22s\nmemcached : Ensuring config directories exist --------------------------- 1.73s\nGroup hosts based on Kolla action --------------------------------------- 0.64s\nmemcached : include_tasks ----------------------------------------------- 0.53s\nGroup hosts based on enabled services ----------------------------------- 0.48s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  00:47:19 +0000 (0:00:00.473)       0:00:00.473 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  00:47:20 +0000 (0:00:00.550)       0:00:01.023 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_redis_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_redis_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_redis_True)\u001b[0m\n\nPLAY [Apply role redis] ********************************************************\n\nTASK [redis : include_tasks] ***************************************************\nSaturday 21 March 2026  00:47:20 +0000 (0:00:00.827)       0:00:01.850 ******** \n\u001b[0;36mincluded: /ansible/roles/redis/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [redis : Ensuring config directories exist] *******************************\nSaturday 21 March 2026  00:47:22 +0000 (0:00:01.288)       0:00:03.139 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\nTASK [redis : Copying over default config.json files] **************************\nSaturday 21 March 2026  00:47:25 +0000 (0:00:02.777)       0:00:05.916 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\nTASK [redis : Copying over redis config files] *********************************\nSaturday 21 March 2026  00:47:28 +0000 (0:00:03.126)       0:00:09.043 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\nTASK [redis : Check redis containers] ******************************************\nSaturday 21 March 2026  00:47:30 +0000 (0:00:02.760)       0:00:11.804 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\nTASK [redis : Flush handlers] **************************************************\nSaturday 21 March 2026  00:47:32 +0000 (0:00:01.625)       0:00:13.429 ******** \n\nTASK [redis : Flush handlers] **************************************************\nSaturday 21 March 2026  00:47:32 +0000 (0:00:00.222)       0:00:13.651 ******** \n\nTASK [redis : Flush handlers] **************************************************\nSaturday 21 March 2026  00:47:32 +0000 (0:00:00.055)       0:00:13.707 ******** \n\nRUNNING HANDLER [redis : Restart redis container] ******************************\nSaturday 21 March 2026  00:47:32 +0000 (0:00:00.117)       0:00:13.825 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [redis : Restart redis-sentinel container] *********************\nSaturday 21 March 2026  00:47:40 +0000 (0:00:07.616)       0:00:21.441 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:47:50 +0000 (0:00:09.991)       0:00:31.433 ******** \n=============================================================================== \nredis : Restart redis-sentinel container -------------------------------- 9.99s\nredis : Restart redis container ----------------------------------------- 7.62s\nredis : Copying over default config.json files -------------------------- 3.13s\nredis : Ensuring config directories exist ------------------------------- 2.78s\nredis : Copying over redis config files --------------------------------- 2.76s\nredis : Check redis containers ------------------------------------------ 1.63s\nredis : include_tasks --------------------------------------------------- 1.29s\nGroup hosts based on enabled services ----------------------------------- 0.83s\nGroup hosts based on Kolla action --------------------------------------- 0.55s\nredis : Flush handlers -------------------------------------------------- 0.40s\n\u001b[32m2026-03-21 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c076440c-db82-4d81-8e70-68ed60306a74 is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-21 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:47:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:47:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:47:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:48:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:48:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:48:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:48:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  00:47:19 +0000 (0:00:00.380)       0:00:00.380 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  00:47:20 +0000 (0:00:01.067)       0:00:01.448 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\nPLAY [Apply role openvswitch] **************************************************\n\nTASK [openvswitch : include_tasks] *********************************************\nSaturday 21 March 2026  00:47:21 +0000 (0:00:01.502)       0:00:02.951 ******** \n\u001b[0;36mincluded: /ansible/roles/openvswitch/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [module-load : Load modules] **********************************************\nSaturday 21 March 2026  00:47:24 +0000 (0:00:02.286)       0:00:05.237 ******** \n\u001b[0;33mchanged: [testbed-node-5] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=openvswitch)\u001b[0m\n\nTASK [module-load : Persist modules via modules-load.d] ************************\nSaturday 21 March 2026  00:47:26 +0000 (0:00:02.369)       0:00:07.607 ******** \n\u001b[0;33mchanged: [testbed-node-5] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=openvswitch)\u001b[0m\n\nTASK [module-load : Drop module persistence] ***********************************\nSaturday 21 March 2026  00:47:28 +0000 (0:00:01.681)       0:00:09.288 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [openvswitch : Create /run/openvswitch directory on host] *****************\nSaturday 21 March 2026  00:47:29 +0000 (0:00:01.706)       0:00:10.995 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [openvswitch : Ensuring config directories exist] *************************\nSaturday 21 March 2026  00:47:30 +0000 (0:00:00.733)       0:00:11.728 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\nTASK [openvswitch : Copying over config.json files for services] ***************\nSaturday 21 March 2026  00:47:32 +0000 (0:00:01.552)       0:00:13.281 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\nTASK [openvswitch : Copying over ovs-vsctl wrapper] ****************************\nSaturday 21 March 2026  00:47:34 +0000 (0:00:02.836)       0:00:16.118 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [openvswitch : Check openvswitch containers] ******************************\nSaturday 21 March 2026  00:47:36 +0000 (0:00:01.424)       0:00:17.543 ******** \n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\nTASK [openvswitch : Flush Handlers] ********************************************\nSaturday 21 March 2026  00:47:39 +0000 (0:00:03.181)       0:00:20.724 ******** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nSaturday 21 March 2026  00:47:39 +0000 (0:00:00.193)       0:00:20.918 ******** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nSaturday 21 March 2026  00:47:39 +0000 (0:00:00.222)       0:00:21.141 ******** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nSaturday 21 March 2026  00:47:40 +0000 (0:00:00.170)       0:00:21.312 ******** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nSaturday 21 March 2026  00:47:40 +0000 (0:00:00.324)       0:00:21.636 ******** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nSaturday 21 March 2026  00:47:40 +0000 (0:00:00.275)       0:00:21.911 ******** \n\nRUNNING HANDLER [openvswitch : Restart openvswitch-db-server container] ********\nSaturday 21 March 2026  00:47:41 +0000 (0:00:00.494)       0:00:22.406 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [openvswitch : Waiting for openvswitch_db service to be ready] ***\nSaturday 21 March 2026  00:47:51 +0000 (0:00:10.009)       0:00:32.416 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [openvswitch : Restart openvswitch-vswitchd container] *********\nSaturday 21 March 2026  00:47:52 +0000 (0:00:01.554)       0:00:33.971 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [openvswitch : Set system-id, hostname and hw-offload] ********************\nSaturday 21 March 2026  00:48:03 +0000 (0:00:10.281)       0:00:44.252 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-3'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-4'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-1'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-5'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-0'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-2'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-3'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-1'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-4'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-0'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-5'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-2'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\nTASK [openvswitch : Ensuring OVS bridge is properly setup] *********************\nSaturday 21 March 2026  00:48:10 +0000 (0:00:07.157)       0:00:51.409 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=br-ex) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=br-ex) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=br-ex) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=br-ex)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=br-ex)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=br-ex)\u001b[0m\n\nTASK [openvswitch : Ensuring OVS ports are properly setup] *********************\nSaturday 21 March 2026  00:48:13 +0000 (0:00:02.927)       0:00:54.336 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=['br-ex', 'vxlan0']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=['br-ex', 'vxlan0']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=['br-ex', 'vxlan0']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=['br-ex', 'vxlan0'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=['br-ex', 'vxlan0'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=['br-ex', 'vxlan0'])\u001b[0m\n\nRUNNING HANDLER [openvswitch : Restart openvswitch-vswitchd container] *********\nSaturday 21 March 2026  00:48:16 +0000 (0:00:03.512)       0:00:57.849 ******** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:48:24 +0000 (0:00:08.015)       0:01:05.865 ******** \n=============================================================================== \nopenvswitch : Restart openvswitch-vswitchd container ------------------- 18.30s\nopenvswitch : Restart openvswitch-db-server container ------------------ 10.01s\nopenvswitch : Set system-id, hostname and hw-offload -------------------- 7.16s\nopenvswitch : Ensuring OVS ports are properly setup --------------------- 3.51s\nopenvswitch : Check openvswitch containers ------------------------------ 3.18s\nopenvswitch : Ensuring OVS bridge is properly setup --------------------- 2.93s\nopenvswitch : Copying over config.json files for services --------------- 2.84s\nmodule-load : Load modules ---------------------------------------------- 2.37s\nopenvswitch : include_tasks --------------------------------------------- 2.29s\nmodule-load : Drop module persistence ----------------------------------- 1.71s\nmodule-load : Persist modules via modules-load.d ------------------------ 1.68s\nopenvswitch : Flush Handlers -------------------------------------------- 1.68s\nopenvswitch : Waiting for openvswitch_db service to be ready ------------ 1.55s\nopenvswitch : Ensuring config directories exist ------------------------- 1.55s\nGroup hosts based on enabled services ----------------------------------- 1.50s\nopenvswitch : Copying over ovs-vsctl wrapper ---------------------------- 1.42s\nGroup hosts based on Kolla action --------------------------------------- 1.07s\nopenvswitch : Create /run/openvswitch directory on host ----------------- 0.73s\n\u001b[32m2026-03-21 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-21 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:48:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:48:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:48:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:48:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:48:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:48:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:48:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:48:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:48:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d647732c-4957-44b7-8e9a-a38699f22e53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\n\u001b[32m2026-03-21 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state SUCCESS\u001b[0m\n\nPLAY [Prepare all k3s nodes] ***************************************************\n\nTASK [k3s_prereq : Validating arguments against arg spec 'main' - Prerequisites] ***\nSaturday 21 March 2026  00:44:37 +0000 (0:00:00.261)       0:00:00.261 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Set same timezone on every Server] **************************\nSaturday 21 March 2026  00:44:38 +0000 (0:00:00.552)       0:00:00.813 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Set SELinux to disabled state] ******************************\nSaturday 21 March 2026  00:44:39 +0000 (0:00:00.759)       0:00:01.572 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Enable IPv4 forwarding] *************************************\nSaturday 21 March 2026  00:44:39 +0000 (0:00:00.590)       0:00:02.163 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Enable IPv6 forwarding] *************************************\nSaturday 21 March 2026  00:44:42 +0000 (0:00:02.271)       0:00:04.434 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Enable IPv6 router advertisements] **************************\nSaturday 21 March 2026  00:44:43 +0000 (0:00:01.025)       0:00:05.460 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Add br_netfilter to /etc/modules-load.d/] *******************\nSaturday 21 March 2026  00:44:44 +0000 (0:00:01.250)       0:00:06.711 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Load br_netfilter] ******************************************\nSaturday 21 March 2026  00:44:45 +0000 (0:00:00.641)       0:00:07.352 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Set bridge-nf-call-iptables (just to be sure)] **************\nSaturday 21 March 2026  00:44:45 +0000 (0:00:00.750)       0:00:08.103 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Add /usr/local/bin to sudo secure_path] *********************\nSaturday 21 March 2026  00:44:46 +0000 (0:00:00.767)       0:00:08.870 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_download : Validating arguments against arg spec 'main' - Manage the downloading of K3S binaries] ***\nSaturday 21 March 2026  00:44:47 +0000 (0:00:00.988)       0:00:09.858 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_download : Download k3s binary x64] **********************************\nSaturday 21 March 2026  00:44:48 +0000 (0:00:01.196)       0:00:11.054 ******** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [k3s_download : Download k3s binary arm64] ********************************\nSaturday 21 March 2026  00:44:55 +0000 (0:00:06.510)       0:00:17.565 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_download : Download k3s binary armhf] ********************************\nSaturday 21 March 2026  00:44:56 +0000 (0:00:01.176)       0:00:18.741 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_custom_registries : Validating arguments against arg spec 'main' - Configure the use of a custom container registry] ***\nSaturday 21 March 2026  00:44:59 +0000 (0:00:02.999)       0:00:21.741 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_custom_registries : Create directory /etc/rancher/k3s] ***************\nSaturday 21 March 2026  00:45:00 +0000 (0:00:01.399)       0:00:23.140 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_custom_registries : Insert registries into /etc/rancher/k3s/registries.yaml] ***\nSaturday 21 March 2026  00:45:01 +0000 (0:00:01.028)       0:00:24.169 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_custom_registries : Remove /etc/rancher/k3s/registries.yaml when no registries configured] ***\nSaturday 21 March 2026  00:45:03 +0000 (0:00:01.311)       0:00:25.481 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Deploy k3s master nodes] *************************************************\n\nTASK [k3s_server : Validating arguments against arg spec 'main' - Setup k3s servers] ***\nSaturday 21 March 2026  00:45:04 +0000 (0:00:01.589)       0:00:27.070 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Stop k3s-init] **********************************************\nSaturday 21 March 2026  00:45:06 +0000 (0:00:01.640)       0:00:28.711 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [k3s_server : Stop k3s] ***************************************************\nSaturday 21 March 2026  00:45:08 +0000 (0:00:02.034)       0:00:30.746 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Clean previous runs of k3s-init] ****************************\nSaturday 21 March 2026  00:45:09 +0000 (0:00:00.978)       0:00:31.724 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Deploy K3s http_proxy conf] *********************************\nSaturday 21 March 2026  00:45:10 +0000 (0:00:01.457)       0:00:33.182 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create /etc/rancher/k3s directory] **************************\nSaturday 21 March 2026  00:45:11 +0000 (0:00:00.695)       0:00:33.877 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create custom resolv.conf for k3s] **************************\nSaturday 21 March 2026  00:45:12 +0000 (0:00:00.936)       0:00:34.813 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Deploy vip manifest] ****************************************\nSaturday 21 March 2026  00:45:14 +0000 (0:00:01.635)       0:00:36.449 ******** \n\u001b[0;36mincluded: /ansible/roles/k3s_server/tasks/vip.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [k3s_server : Set _kube_vip_bgp_peers fact] *******************************\nSaturday 21 March 2026  00:45:14 +0000 (0:00:00.646)       0:00:37.096 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create manifests directory on first master] *****************\nSaturday 21 March 2026  00:45:16 +0000 (0:00:02.140)       0:00:39.236 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Download vip rbac manifest to first master] *****************\nSaturday 21 March 2026  00:45:17 +0000 (0:00:00.599)       0:00:39.836 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Copy vip manifest to first master] **************************\nSaturday 21 March 2026  00:45:18 +0000 (0:00:01.186)       0:00:41.022 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Deploy metallb manifest] ************************************\nSaturday 21 March 2026  00:45:20 +0000 (0:00:01.319)       0:00:42.342 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Deploy kube-vip manifest] ***********************************\nSaturday 21 March 2026  00:45:20 +0000 (0:00:00.638)       0:00:42.981 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Init cluster inside the transient k3s-init service] *********\nSaturday 21 March 2026  00:45:21 +0000 (0:00:00.380)       0:00:43.361 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [k3s_server : Detect Kubernetes version for label compatibility] **********\nSaturday 21 March 2026  00:45:24 +0000 (0:00:03.428)       0:00:46.791 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [k3s_server : Set node role label selector based on Kubernetes version] ***\nSaturday 21 March 2026  00:45:27 +0000 (0:00:02.672)       0:00:49.464 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails)] ***\nSaturday 21 March 2026  00:45:28 +0000 (0:00:00.869)       0:00:50.334 ******** \n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Save logs of k3s-init.service] ******************************\nSaturday 21 March 2026  00:46:11 +0000 (0:00:43.380)       0:01:33.714 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Kill the temporary service used for initialization] *********\nSaturday 21 March 2026  00:46:11 +0000 (0:00:00.435)       0:01:34.149 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Copy K3s service file] **************************************\nSaturday 21 March 2026  00:46:12 +0000 (0:00:00.845)       0:01:34.995 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Enable and check K3s service] *******************************\nSaturday 21 March 2026  00:46:14 +0000 (0:00:01.401)       0:01:36.396 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Wait for node-token] ****************************************\nSaturday 21 March 2026  00:46:39 +0000 (0:00:25.365)       0:02:01.762 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Register node-token file access mode] ***********************\nSaturday 21 March 2026  00:46:40 +0000 (0:00:00.714)       0:02:02.477 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Change file access node-token] ******************************\nSaturday 21 March 2026  00:46:42 +0000 (0:00:01.914)       0:02:04.391 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Read node-token from master] ********************************\nSaturday 21 March 2026  00:46:42 +0000 (0:00:00.702)       0:02:05.094 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Store Master node-token] ************************************\nSaturday 21 March 2026  00:46:43 +0000 (0:00:00.630)       0:02:05.725 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Restore node-token file access] *****************************\nSaturday 21 March 2026  00:46:43 +0000 (0:00:00.541)       0:02:06.266 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create directory .kube] *************************************\nSaturday 21 March 2026  00:46:44 +0000 (0:00:00.968)       0:02:07.235 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Copy config file to user home directory] ********************\nSaturday 21 March 2026  00:46:45 +0000 (0:00:00.808)       0:02:08.043 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Configure kubectl cluster to https://192.168.16.8:6443] *****\nSaturday 21 March 2026  00:46:46 +0000 (0:00:00.872)       0:02:08.915 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create kubectl symlink] *************************************\nSaturday 21 March 2026  00:46:47 +0000 (0:00:00.827)       0:02:09.743 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create crictl symlink] **************************************\nSaturday 21 March 2026  00:46:47 +0000 (0:00:00.415)       0:02:10.158 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Get contents of manifests folder] ***************************\nSaturday 21 March 2026  00:46:48 +0000 (0:00:00.263)       0:02:10.422 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Get sub dirs of manifests folder] ***************************\nSaturday 21 March 2026  00:46:48 +0000 (0:00:00.618)       0:02:11.040 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start] ***\nSaturday 21 March 2026  00:46:49 +0000 (0:00:00.623)       0:02:11.664 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip-rbac.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m\n\nPLAY [Deploy k3s worker nodes] *************************************************\n\nTASK [k3s_agent : Validating arguments against arg spec 'main' - Setup k3s agents] ***\nSaturday 21 March 2026  00:46:52 +0000 (0:00:02.785)       0:02:14.449 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Check if system is PXE-booted] *******************************\nSaturday 21 March 2026  00:46:52 +0000 (0:00:00.311)       0:02:14.760 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Set fact for PXE-booted system] ******************************\nSaturday 21 March 2026  00:46:53 +0000 (0:00:00.553)       0:02:15.314 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Include http_proxy configuration tasks] **********************\nSaturday 21 March 2026  00:46:53 +0000 (0:00:00.463)       0:02:15.778 ******** \n\u001b[0;36mincluded: /ansible/roles/k3s_agent/tasks/http_proxy.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [k3s_agent : Create k3s-node.service.d directory] *************************\nSaturday 21 March 2026  00:46:53 +0000 (0:00:00.449)       0:02:16.227 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Copy K3s http_proxy conf file] *******************************\nSaturday 21 March 2026  00:46:54 +0000 (0:00:00.251)       0:02:16.479 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Deploy K3s http_proxy conf] **********************************\nSaturday 21 March 2026  00:46:54 +0000 (0:00:00.386)       0:02:16.865 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Create /etc/rancher/k3s directory] ***************************\nSaturday 21 March 2026  00:46:54 +0000 (0:00:00.294)       0:02:17.159 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Create custom resolv.conf for k3s] ***************************\nSaturday 21 March 2026  00:46:55 +0000 (0:00:00.650)       0:02:17.810 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Configure the k3s service] ***********************************\nSaturday 21 March 2026  00:46:56 +0000 (0:00:00.959)       0:02:18.769 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Manage k3s service] ******************************************\nSaturday 21 March 2026  00:46:58 +0000 (0:00:01.682)       0:02:20.452 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY [Prepare kubeconfig file] *************************************************\n\nTASK [Get home directory of operator user] *************************************\nSaturday 21 March 2026  00:47:08 +0000 (0:00:10.659)       0:02:31.112 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Create .kube directory] **************************************************\nSaturday 21 March 2026  00:47:09 +0000 (0:00:00.785)       0:02:31.898 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Get kubeconfig file] *****************************************************\nSaturday 21 March 2026  00:47:10 +0000 (0:00:00.410)       0:02:32.309 ******** \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [Write kubeconfig file] ***************************************************\nSaturday 21 March 2026  00:47:10 +0000 (0:00:00.543)       0:02:32.852 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Change server address in the kubeconfig] *********************************\nSaturday 21 March 2026  00:47:11 +0000 (0:00:00.988)       0:02:33.841 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Make kubeconfig available for use inside the manager service] ************\nSaturday 21 March 2026  00:47:12 +0000 (0:00:00.522)       0:02:34.363 ******** \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Change server address in the kubeconfig inside the manager service] ******\nSaturday 21 March 2026  00:47:13 +0000 (0:00:01.688)       0:02:36.052 ******** \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Set KUBECONFIG environment variable] *************************************\nSaturday 21 March 2026  00:47:14 +0000 (0:00:00.897)       0:02:36.950 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Enable kubectl command line completion] **********************************\nSaturday 21 March 2026  00:47:15 +0000 (0:00:00.394)       0:02:37.344 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY [Apply role kubectl] ******************************************************\n\nTASK [kubectl : Gather variables for each operating system] ********************\nSaturday 21 March 2026  00:47:15 +0000 (0:00:00.395)       0:02:37.739 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Include distribution specific install tasks] *******************\nSaturday 21 March 2026  00:47:15 +0000 (0:00:00.127)       0:02:37.867 ******** \n\u001b[0;36mincluded: /ansible/roles/kubectl/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [kubectl : Remove old architecture-dependent repository] ******************\nSaturday 21 March 2026  00:47:15 +0000 (0:00:00.202)       0:02:38.070 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Install apt-transport-https package] ***************************\nSaturday 21 March 2026  00:47:16 +0000 (0:00:01.053)       0:02:39.123 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Add repository gpg key] ****************************************\nSaturday 21 March 2026  00:47:18 +0000 (0:00:01.300)       0:02:40.424 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Set permissions of gpg key] ************************************\nSaturday 21 March 2026  00:47:19 +0000 (0:00:00.958)       0:02:41.383 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Add repository Debian] *****************************************\nSaturday 21 March 2026  00:47:19 +0000 (0:00:00.503)       0:02:41.887 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Install required packages] *************************************\nSaturday 21 March 2026  00:47:26 +0000 (0:00:06.754)       0:02:48.641 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Remove kubectl symlink] ****************************************\nSaturday 21 March 2026  00:47:40 +0000 (0:00:14.199)       0:03:02.841 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY [Run post actions on master nodes] ****************************************\n\nTASK [k3s_server_post : Validating arguments against arg spec 'main' - Configure k3s cluster] ***\nSaturday 21 March 2026  00:47:41 +0000 (0:00:00.525)       0:03:03.366 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server_post : Deploy calico] *****************************************\nSaturday 21 March 2026  00:47:41 +0000 (0:00:00.502)       0:03:03.868 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server_post : Deploy cilium] *****************************************\nSaturday 21 March 2026  00:47:41 +0000 (0:00:00.342)       0:03:04.211 ******** \n\u001b[0;36mincluded: /ansible/roles/k3s_server_post/tasks/cilium.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [k3s_server_post : Create tmp directory on first master] ******************\nSaturday 21 March 2026  00:47:42 +0000 (0:00:00.572)       0:03:04.783 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Wait for connectivity to kube VIP] *********************\nSaturday 21 March 2026  00:47:43 +0000 (0:00:00.853)       0:03:05.637 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Fail if kube VIP not reachable] ************************\nSaturday 21 March 2026  00:47:44 +0000 (0:00:00.938)       0:03:06.576 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Test for existing Cilium install] **********************\nSaturday 21 March 2026  00:47:44 +0000 (0:00:00.282)       0:03:06.858 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Check Cilium version] **********************************\nSaturday 21 March 2026  00:47:45 +0000 (0:00:01.059)       0:03:07.918 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Parse installed Cilium version] ************************\nSaturday 21 March 2026  00:47:45 +0000 (0:00:00.101)       0:03:08.020 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Determine if Cilium needs update] **********************\nSaturday 21 March 2026  00:47:45 +0000 (0:00:00.122)       0:03:08.142 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Log result] ********************************************\nSaturday 21 March 2026  00:47:45 +0000 (0:00:00.105)       0:03:08.247 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Install Cilium] ****************************************\nSaturday 21 March 2026  00:47:46 +0000 (0:00:00.112)       0:03:08.360 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Wait for Cilium resources] *****************************\nSaturday 21 March 2026  00:47:50 +0000 (0:00:04.198)       0:03:12.558 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/cilium-operator)\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0 -> localhost]: Wait for Cilium resources (30 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=daemonset/cilium)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/hubble-relay)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/hubble-ui)\u001b[0m\n\nTASK [k3s_server_post : Set _cilium_bgp_neighbors fact] ************************\nSaturday 21 March 2026  00:48:33 +0000 (0:00:42.773)       0:03:55.332 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Copy BGP manifests to first master] ********************\nSaturday 21 March 2026  00:48:34 +0000 (0:00:01.129)       0:03:56.462 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Apply BGP manifests] ***********************************\nSaturday 21 March 2026  00:48:35 +0000 (0:00:01.569)       0:03:58.031 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Print error message if BGP manifests application fails] ***\nSaturday 21 March 2026  00:48:36 +0000 (0:00:00.952)       0:03:58.983 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Test for BGP config resources] *************************\nSaturday 21 March 2026  00:48:36 +0000 (0:00:00.114)       0:03:59.097 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=kubectl get CiliumBGPPeeringPolicy.cilium.io)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=kubectl get CiliumLoadBalancerIPPool.cilium.io)\u001b[0m\n\nTASK [k3s_server_post : Deploy metallb pool] ***********************************\nSaturday 21 March 2026  00:48:38 +0000 (0:00:01.792)       0:04:00.889 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server_post : Remove tmp directory used for manifests] ***************\nSaturday 21 March 2026  00:48:38 +0000 (0:00:00.319)       0:04:01.209 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role k9s] **********************************************************\n\nTASK [k9s : Gather variables for each operating system] ************************\nSaturday 21 March 2026  00:48:40 +0000 (0:00:01.128)       0:04:02.337 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [k9s : Include distribution specific install tasks] ***********************\nSaturday 21 March 2026  00:48:40 +0000 (0:00:00.169)       0:04:02.507 ******** \n\u001b[0;36mincluded: /ansible/roles/k9s/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [k9s : Install k9s packages] **********************************************\nSaturday 21 March 2026  00:48:40 +0000 (0:00:00.412)       0:04:02.920 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY [Manage labels, annotations, and taints on all k3s nodes] *****************\n\nTASK [Merge labels, annotations, and taints] ***********************************\nSaturday 21 March 2026  00:48:46 +0000 (0:00:05.891)       0:04:08.812 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Manage labels] ***********************************************************\nSaturday 21 March 2026  00:48:47 +0000 (0:00:00.606)       0:04:09.418 ******** \n\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m\n\nTASK [Manage annotations] ******************************************************\nSaturday 21 March 2026  00:48:58 +0000 (0:00:11.474)       0:04:20.892 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Manage taints] ***********************************************************\nSaturday 21 March 2026  00:48:59 +0000 (0:00:00.454)       0:04:21.347 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=50  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=28  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:48:59 +0000 (0:00:00.526)       0:04:21.874 ******** \n=============================================================================== \nk3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails) -- 43.38s\nk3s_server_post : Wait for Cilium resources ---------------------------- 42.77s\nk3s_server : Enable and check K3s service ------------------------------ 25.37s\nkubectl : Install required packages ------------------------------------ 14.20s\nManage labels ---------------------------------------------------------- 11.47s\nk3s_agent : Manage k3s service ----------------------------------------- 10.66s\nkubectl : Add repository Debian ----------------------------------------- 6.75s\nk3s_download : Download k3s binary x64 ---------------------------------- 6.51s\nk9s : Install k9s packages ---------------------------------------------- 5.89s\nk3s_server_post : Install Cilium ---------------------------------------- 4.20s\nk3s_server : Init cluster inside the transient k3s-init service --------- 3.43s\nk3s_download : Download k3s binary armhf -------------------------------- 3.00s\nk3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start --- 2.79s\nk3s_server : Detect Kubernetes version for label compatibility ---------- 2.67s\nk3s_prereq : Enable IPv4 forwarding ------------------------------------- 2.27s\nk3s_server : Set _kube_vip_bgp_peers fact ------------------------------- 2.14s\nk3s_server : Stop k3s-init ---------------------------------------------- 2.03s\nk3s_server : Register node-token file access mode ----------------------- 1.91s\nk3s_server_post : Test for BGP config resources ------------------------- 1.79s\nMake kubeconfig available for use inside the manager service ------------ 1.69s\n\u001b[32m2026-03-21 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 591381dc-15c5-47b5-8a6a-1c0023f4941b is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d647732c-4957-44b7-8e9a-a38699f22e53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 591381dc-15c5-47b5-8a6a-1c0023f4941b is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:49:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d647732c-4957-44b7-8e9a-a38699f22e53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 591381dc-15c5-47b5-8a6a-1c0023f4941b is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-21 00:49:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:49:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d647732c-4957-44b7-8e9a-a38699f22e53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:49:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d647732c-4957-44b7-8e9a-a38699f22e53 is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-21 00:49:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:49:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:49:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:49:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:49:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:49:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:49:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:49:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:49:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:49:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:49:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:49:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state SUCCESS\u001b[0m\n\n\nPLAY [Copy kubeconfig to the configuration repository] *************************\n\nTASK [Get kubeconfig file] *****************************************************\nSaturday 21 March 2026  00:49:02 +0000 (0:00:00.230)       0:00:00.230 ******** \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [Write kubeconfig file] ***************************************************\nSaturday 21 March 2026  00:49:04 +0000 (0:00:01.458)       0:00:01.688 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Change server address in the kubeconfig file] ****************************\nSaturday 21 March 2026  00:49:06 +0000 (0:00:01.658)       0:00:03.347 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:49:06 +0000 (0:00:00.707)       0:00:04.055 ******** \n=============================================================================== \nWrite kubeconfig file --------------------------------------------------- 1.66s\nGet kubeconfig file ----------------------------------------------------- 1.46s\nChange server address in the kubeconfig file ---------------------------- 0.71s\n\n\nPLAY [Prepare kubeconfig file] *************************************************\n\nTASK [Get home directory of operator user] *************************************\nSaturday 21 March 2026  00:49:02 +0000 (0:00:00.225)       0:00:00.225 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Create .kube directory] **************************************************\nSaturday 21 March 2026  00:49:03 +0000 (0:00:00.764)       0:00:00.990 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Get kubeconfig file] *****************************************************\nSaturday 21 March 2026  00:49:03 +0000 (0:00:00.495)       0:00:01.485 ******** \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [Write kubeconfig file] ***************************************************\nSaturday 21 March 2026  00:49:04 +0000 (0:00:00.985)       0:00:02.471 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Change server address in the kubeconfig] *********************************\nSaturday 21 March 2026  00:49:05 +0000 (0:00:01.111)       0:00:03.582 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Make kubeconfig available for use inside the manager service] ************\nSaturday 21 March 2026  00:49:06 +0000 (0:00:00.732)       0:00:04.314 ******** \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Change server address in the kubeconfig inside the manager service] ******\nSaturday 21 March 2026  00:49:08 +0000 (0:00:01.848)       0:00:06.163 ******** \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Set KUBECONFIG environment variable] *************************************\nSaturday 21 March 2026  00:49:09 +0000 (0:00:01.482)       0:00:07.646 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Enable kubectl command line completion] **********************************\nSaturday 21 March 2026  00:49:10 +0000 (0:00:00.518)       0:00:08.165 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:49:10 +0000 (0:00:00.528)       0:00:08.693 ******** \n=============================================================================== \nMake kubeconfig available for use inside the manager service ------------ 1.85s\nChange server address in the kubeconfig inside the manager service ------ 1.48s\nWrite kubeconfig file --------------------------------------------------- 1.11s\nGet kubeconfig file ----------------------------------------------------- 0.99s\nGet home directory of operator user ------------------------------------- 0.76s\nChange server address in the kubeconfig --------------------------------- 0.73s\nEnable kubectl command line completion ---------------------------------- 0.53s\nSet KUBECONFIG environment variable ------------------------------------- 0.52s\nCreate .kube directory -------------------------------------------------- 0.50s\n\n\nPLAY [Set kolla_action_rabbitmq] ***********************************************\n\nTASK [Inform the user about the following task] ********************************\nSaturday 21 March 2026  00:47:40 +0000 (0:00:00.096)       0:00:00.096 ******** \n\u001b[0;32mok: [localhost] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"The task 'Check RabbitMQ service' fails if the RabbitMQ service has not yet been deployed. This is fine.\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Check RabbitMQ service] **************************************************\nSaturday 21 March 2026  00:47:40 +0000 (0:00:00.048)       0:00:00.145 ******** \n\u001b[0;31mfatal: [localhost]: FAILED! => {\"changed\": false, \"elapsed\": 2, \"msg\": \"Timeout when waiting for search string RabbitMQ Management in 192.168.16.9:15672\"}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\nTASK [Set kolla_action_rabbitmq = upgrade if RabbitMQ is already running] ******\nSaturday 21 March 2026  00:47:45 +0000 (0:00:04.754)       0:00:04.899 ******** \n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nTASK [Set kolla_action_rabbitmq = kolla_action_ng] *****************************\nSaturday 21 March 2026  00:47:45 +0000 (0:00:00.063)       0:00:04.963 ******** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  00:47:45 +0000 (0:00:00.234)       0:00:05.198 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  00:47:45 +0000 (0:00:00.343)       0:00:05.541 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_rabbitmq_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_rabbitmq_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_rabbitmq_True)\u001b[0m\n\nPLAY [Apply role rabbitmq] *****************************************************\n\nTASK [rabbitmq : include_tasks] ************************************************\nSaturday 21 March 2026  00:47:46 +0000 (0:00:00.402)       0:00:05.944 ******** \n\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [rabbitmq : Get container facts] ******************************************\nSaturday 21 March 2026  00:47:46 +0000 (0:00:00.563)       0:00:06.508 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Get current RabbitMQ version] *********************************\nSaturday 21 March 2026  00:47:48 +0000 (0:00:01.788)       0:00:08.297 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Get new RabbitMQ version] *************************************\nSaturday 21 March 2026  00:47:48 +0000 (0:00:00.399)       0:00:08.696 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Check if running RabbitMQ is at most one version behind] ******\nSaturday 21 March 2026  00:47:49 +0000 (0:00:00.394)       0:00:09.090 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Catch when RabbitMQ is being downgraded] **********************\nSaturday 21 March 2026  00:47:49 +0000 (0:00:00.375)       0:00:09.466 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : include_tasks] ************************************************\nSaturday 21 March 2026  00:47:50 +0000 (0:00:00.475)       0:00:09.941 ******** \n\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/remove-ha-all-policy.yml for testbed-node-1, testbed-node-0, testbed-node-2\u001b[0m\n\nTASK [rabbitmq : Get container facts] ******************************************\nSaturday 21 March 2026  00:47:51 +0000 (0:00:01.233)       0:00:11.175 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : List RabbitMQ policies] ***************************************\nSaturday 21 March 2026  00:47:52 +0000 (0:00:01.319)       0:00:12.495 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Remove ha-all policy from RabbitMQ] ***************************\nSaturday 21 March 2026  00:47:54 +0000 (0:00:01.861)       0:00:14.356 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Ensuring config directories exist] ****************************\nSaturday 21 March 2026  00:47:55 +0000 (0:00:00.652)       0:00:15.009 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [rabbitmq : Copying over config.json files for services] ******************\nSaturday 21 March 2026  00:47:56 +0000 (0:00:01.202)       0:00:16.212 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [rabbitmq : Copying over rabbitmq-env.conf] *******************************\nSaturday 21 March 2026  00:47:57 +0000 (0:00:01.557)       0:00:17.770 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over rabbitmq.conf] ***********************************\nSaturday 21 March 2026  00:47:59 +0000 (0:00:01.801)       0:00:19.572 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over erl_inetrc] **************************************\nSaturday 21 March 2026  00:48:02 +0000 (0:00:02.319)       0:00:21.892 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over advanced.config] *********************************\nSaturday 21 March 2026  00:48:03 +0000 (0:00:01.216)       0:00:23.108 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over definitions.json] ********************************\nSaturday 21 March 2026  00:48:04 +0000 (0:00:01.681)       0:00:24.790 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over enabled_plugins] *********************************\nSaturday 21 March 2026  00:48:06 +0000 (0:00:01.400)       0:00:26.190 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m\n\nTASK [rabbitmq : include_tasks] ************************************************\nSaturday 21 March 2026  00:48:08 +0000 (0:00:01.813)       0:00:28.004 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Check rabbitmq containers] ************************************\nSaturday 21 March 2026  00:48:08 +0000 (0:00:00.397)       0:00:28.402 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [rabbitmq : Creating rabbitmq volume] *************************************\nSaturday 21 March 2026  00:48:09 +0000 (0:00:01.001)       0:00:29.403 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Running RabbitMQ bootstrap container] *************************\nSaturday 21 March 2026  00:48:10 +0000 (0:00:00.821)       0:00:30.224 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [rabbitmq : Restart rabbitmq container] ************************\nSaturday 21 March 2026  00:48:16 +0000 (0:00:06.209)       0:00:36.434 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Restart rabbitmq services] ***********************************************\n\nTASK [rabbitmq : Get info on RabbitMQ container] *******************************\nSaturday 21 March 2026  00:48:16 +0000 (0:00:00.349)       0:00:36.784 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************\nSaturday 21 March 2026  00:48:17 +0000 (0:00:00.773)       0:00:37.557 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Restart rabbitmq container] ***********************************\nSaturday 21 March 2026  00:48:18 +0000 (0:00:00.600)       0:00:38.158 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Waiting for rabbitmq to start] ********************************\nSaturday 21 March 2026  00:48:20 +0000 (0:00:01.989)       0:00:40.148 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY [Restart rabbitmq services] ***********************************************\n\nTASK [rabbitmq : Get info on RabbitMQ container] *******************************\nSaturday 21 March 2026  00:49:13 +0000 (0:00:53.364)       0:01:33.512 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************\nSaturday 21 March 2026  00:49:14 +0000 (0:00:00.537)       0:01:34.050 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [rabbitmq : Restart rabbitmq container] ***********************************\nSaturday 21 March 2026  00:49:14 +0000 (0:00:00.225)       0:01:34.276 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [rabbitmq : Waiting for rabbitmq to start] ********************************\nSaturday 21 March 2026  00:49:16 +0000 (0:00:02.332)       0:01:36.609 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY [Restart rabbitmq services] ***********************************************\n\nTASK [rabbitmq : Get info on RabbitMQ container] *******************************\nSaturday 21 March 2026  00:49:29 +0000 (0:00:13.207)       0:01:49.817 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************\nSaturday 21 March 2026  00:49:30 +0000 (0:00:00.589)       0:01:50.406 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Restart rabbitmq container] ***********************************\nSaturday 21 March 2026  00:49:30 +0000 (0:00:00.217)       0:01:50.624 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Waiting for rabbitmq to start] ********************************\nSaturday 21 March 2026  00:49:37 +0000 (0:00:06.464)       0:01:57.088 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Apply rabbitmq post-configuration] ***************************************\n\nTASK [Include rabbitmq post-deploy.yml] ****************************************\nSaturday 21 March 2026  00:49:48 +0000 (0:00:10.785)       0:02:07.874 ******** \n\u001b[0;36mincluded: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [rabbitmq : Enable all stable feature flags] ******************************\nSaturday 21 March 2026  00:49:48 +0000 (0:00:00.663)       0:02:08.537 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m\n\u001b[1;35menable_outward_rabbitmq_True\u001b[0m\n\nPLAY [Apply role rabbitmq (outward)] *******************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m\n\u001b[1;35moutward_rabbitmq_restart\u001b[0m\n\nPLAY [Restart rabbitmq (outward) services] *************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Apply rabbitmq (outward) post-configuration] *****************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mlocalhost\u001b[0m                  : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:49:51 +0000 (0:00:02.529)       0:02:11.066 ******** \n=============================================================================== \nrabbitmq : Waiting for rabbitmq to start ------------------------------- 77.36s\nrabbitmq : Restart rabbitmq container ---------------------------------- 10.79s\nrabbitmq : Running RabbitMQ bootstrap container ------------------------- 6.21s\nCheck RabbitMQ service -------------------------------------------------- 4.75s\nrabbitmq : Enable all stable feature flags ------------------------------ 2.53s\nrabbitmq : Copying over rabbitmq.conf ----------------------------------- 2.32s\nrabbitmq : Get info on RabbitMQ container ------------------------------- 1.90s\nrabbitmq : List RabbitMQ policies --------------------------------------- 1.86s\nrabbitmq : Copying over enabled_plugins --------------------------------- 1.81s\nrabbitmq : Copying over rabbitmq-env.conf ------------------------------- 1.80s\nrabbitmq : Get container facts ------------------------------------------ 1.79s\nrabbitmq : Copying over advanced.config --------------------------------- 1.68s\nrabbitmq : Copying over config.json files for services ------------------ 1.56s\nrabbitmq : Copying over definitions.json -------------------------------- 1.40s\nrabbitmq : Get container facts ------------------------------------------ 1.32s\nrabbitmq : include_tasks ------------------------------------------------ 1.23s\nrabbitmq : Copying over erl_inetrc -------------------------------------- 1.22s\nrabbitmq : Ensuring config directories exist ---------------------------- 1.20s\nrabbitmq : Put RabbitMQ node into maintenance mode ---------------------- 1.04s\nrabbitmq : Check rabbitmq containers ------------------------------------ 1.00s\n\u001b[32m2026-03-21 00:49:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:49:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:49:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:49:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:50:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:50:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:50:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:50:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:50:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:50:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:50:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:50:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:50:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:50:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:50:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  00:48:28 +0000 (0:00:00.161)       0:00:00.161 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  00:48:29 +0000 (0:00:00.799)       0:00:00.961 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_ovn_True)\u001b[0m\n\nPLAY [Apply role ovn-controller] ***********************************************\n\nTASK [ovn-controller : include_tasks] ******************************************\nSaturday 21 March 2026  00:48:30 +0000 (0:00:01.358)       0:00:02.320 ******** \n\u001b[0;36mincluded: /ansible/roles/ovn-controller/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ovn-controller : Ensuring config directories exist] **********************\nSaturday 21 March 2026  00:48:31 +0000 (0:00:01.116)       0:00:03.436 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Copying over config.json files for services] ************\nSaturday 21 March 2026  00:48:33 +0000 (0:00:01.531)       0:00:04.967 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Ensuring systemd override directory exists] *************\nSaturday 21 March 2026  00:48:35 +0000 (0:00:01.747)       0:00:06.715 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Copying over systemd override] **************************\nSaturday 21 March 2026  00:48:37 +0000 (0:00:01.955)       0:00:08.670 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Check ovn-controller containers] ************************\nSaturday 21 March 2026  00:48:38 +0000 (0:00:01.573)       0:00:10.244 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Create br-int bridge on OpenvSwitch] ********************\nSaturday 21 March 2026  00:48:40 +0000 (0:00:02.032)       0:00:12.276 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ovn-controller : Configure OVN in OVSDB] *********************************\nSaturday 21 March 2026  00:48:43 +0000 (0:00:02.796)       0:00:15.073 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.13'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.12'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.14'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.10'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.15'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.11'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:89:18:56', 'state': 'present'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:29:4a:9b', 'state': 'absent'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:2f:fa:44', 'state': 'present'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:52:c1:40', 'state': 'absent'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:71:3a:c3', 'state': 'present'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:33:12:50', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m\n\nTASK [ovn-controller : Flush handlers] *****************************************\nSaturday 21 March 2026  00:49:02 +0000 (0:00:18.425)       0:00:33.499 ******** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nSaturday 21 March 2026  00:49:02 +0000 (0:00:00.090)       0:00:33.589 ******** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nSaturday 21 March 2026  00:49:02 +0000 (0:00:00.067)       0:00:33.656 ******** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nSaturday 21 March 2026  00:49:02 +0000 (0:00:00.064)       0:00:33.721 ******** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nSaturday 21 March 2026  00:49:02 +0000 (0:00:00.058)       0:00:33.780 ******** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nSaturday 21 March 2026  00:49:02 +0000 (0:00:00.071)       0:00:33.851 ******** \n\nRUNNING HANDLER [ovn-controller : Reload systemd config] ***********************\nSaturday 21 March 2026  00:49:02 +0000 (0:00:00.065)       0:00:33.917 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ovn-controller : Restart ovn-controller container] ************\nSaturday 21 March 2026  00:49:04 +0000 (0:00:01.643)       0:00:35.561 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nPLAY [Apply role ovn-db] *******************************************************\n\nTASK [ovn-db : include_tasks] **************************************************\nSaturday 21 March 2026  00:49:31 +0000 (0:00:27.405)       0:01:02.967 ******** \n\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ovn-db : include_tasks] **************************************************\nSaturday 21 March 2026  00:49:31 +0000 (0:00:00.446)       0:01:03.413 ******** \n\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/lookup_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ovn-db : Checking for any existing OVN DB container volumes] *************\nSaturday 21 March 2026  00:49:32 +0000 (0:00:00.661)       0:01:04.075 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN NB volume availability] ***************\nSaturday 21 March 2026  00:49:33 +0000 (0:00:00.999)       0:01:05.075 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN SB volume availability] ***************\nSaturday 21 March 2026  00:49:33 +0000 (0:00:00.331)       0:01:05.407 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Establish whether the OVN NB cluster has already existed] *******\nSaturday 21 March 2026  00:49:34 +0000 (0:00:00.490)       0:01:05.897 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Establish whether the OVN SB cluster has already existed] *******\nSaturday 21 March 2026  00:49:34 +0000 (0:00:00.310)       0:01:06.208 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check if running on all OVN NB DB hosts] ************************\nSaturday 21 March 2026  00:49:35 +0000 (0:00:00.292)       0:01:06.501 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check OVN NB service port liveness] *****************************\nSaturday 21 March 2026  00:49:35 +0000 (0:00:00.289)       0:01:06.790 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN NB service port liveness] *************\nSaturday 21 March 2026  00:49:35 +0000 (0:00:00.296)       0:01:07.086 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Get OVN NB database information] ********************************\nSaturday 21 March 2026  00:49:36 +0000 (0:00:00.486)       0:01:07.573 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN NB leader/follower role] **************\nSaturday 21 March 2026  00:49:36 +0000 (0:00:00.284)       0:01:07.857 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Fail on existing OVN NB cluster with no leader] *****************\nSaturday 21 March 2026  00:49:36 +0000 (0:00:00.268)       0:01:08.126 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check if running on all OVN SB DB hosts] ************************\nSaturday 21 March 2026  00:49:36 +0000 (0:00:00.270)       0:01:08.397 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check OVN SB service port liveness] *****************************\nSaturday 21 March 2026  00:49:37 +0000 (0:00:00.427)       0:01:08.824 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN SB service port liveness] *************\nSaturday 21 March 2026  00:49:37 +0000 (0:00:00.253)       0:01:09.078 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Get OVN SB database information] ********************************\nSaturday 21 March 2026  00:49:37 +0000 (0:00:00.286)       0:01:09.364 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN SB leader/follower role] **************\nSaturday 21 March 2026  00:49:38 +0000 (0:00:00.254)       0:01:09.618 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Fail on existing OVN SB cluster with no leader] *****************\nSaturday 21 March 2026  00:49:38 +0000 (0:00:00.392)       0:01:10.011 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : include_tasks] **************************************************\nSaturday 21 March 2026  00:49:38 +0000 (0:00:00.248)       0:01:10.259 ******** \n\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/bootstrap-initial.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ovn-db : Set bootstrap args fact for NB (new cluster)] *******************\nSaturday 21 March 2026  00:49:39 +0000 (0:00:00.503)       0:01:10.763 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Set bootstrap args fact for SB (new cluster)] *******************\nSaturday 21 March 2026  00:49:39 +0000 (0:00:00.630)       0:01:11.394 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check NB cluster status] ****************************************\nSaturday 21 March 2026  00:49:40 +0000 (0:00:00.419)       0:01:11.814 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check SB cluster status] ****************************************\nSaturday 21 March 2026  00:49:40 +0000 (0:00:00.341)       0:01:12.155 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Remove an old node with the same ip address as the new node in NB DB] ***\nSaturday 21 March 2026  00:49:41 +0000 (0:00:00.407)       0:01:12.563 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Remove an old node with the same ip address as the new node in SB DB] ***\nSaturday 21 March 2026  00:49:41 +0000 (0:00:00.483)       0:01:13.046 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Set bootstrap args fact for NB (new member)] ********************\nSaturday 21 March 2026  00:49:41 +0000 (0:00:00.347)       0:01:13.393 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Set bootstrap args fact for SB (new member)] ********************\nSaturday 21 March 2026  00:49:42 +0000 (0:00:00.269)       0:01:13.663 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Ensuring config directories exist] ******************************\nSaturday 21 March 2026  00:49:42 +0000 (0:00:00.285)       0:01:13.948 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-db : Copying over config.json files for services] ********************\nSaturday 21 March 2026  00:49:44 +0000 (0:00:01.587)       0:01:15.535 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-db : Check ovn containers] *******************************************\nSaturday 21 March 2026  00:49:48 +0000 (0:00:04.183)       0:01:19.719 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-db : Flush handlers] *************************************************\nSaturday 21 March 2026  00:49:50 +0000 (0:00:02.413)       0:01:22.133 ******** \n\nTASK [ovn-db : Flush handlers] *************************************************\nSaturday 21 March 2026  00:49:50 +0000 (0:00:00.122)       0:01:22.255 ******** \n\nTASK [ovn-db : Flush handlers] *************************************************\nSaturday 21 March 2026  00:49:50 +0000 (0:00:00.120)       0:01:22.376 ******** \n\nRUNNING HANDLER [ovn-db : Restart ovn-nb-db container] *************************\nSaturday 21 March 2026  00:49:51 +0000 (0:00:00.134)       0:01:22.511 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ovn-db : Restart ovn-sb-db container] *************************\nSaturday 21 March 2026  00:49:53 +0000 (0:00:02.712)       0:01:25.224 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [ovn-db : Restart ovn-northd container] ************************\nSaturday 21 March 2026  00:50:01 +0000 (0:00:07.696)       0:01:32.920 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ovn-db : Wait for leader election] ***************************************\nSaturday 21 March 2026  00:50:08 +0000 (0:00:07.540)       0:01:40.461 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Get OVN_Northbound cluster leader] ******************************\nSaturday 21 March 2026  00:50:09 +0000 (0:00:00.109)       0:01:40.571 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Configure OVN NB connection settings] ***************************\nSaturday 21 March 2026  00:50:11 +0000 (0:00:01.952)       0:01:42.523 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Get OVN_Southbound cluster leader] ******************************\nSaturday 21 March 2026  00:50:11 +0000 (0:00:00.745)       0:01:43.269 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Configure OVN SB connection settings] ***************************\nSaturday 21 March 2026  00:50:12 +0000 (0:00:01.058)       0:01:44.328 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-nb-db] *********************************************\nSaturday 21 March 2026  00:50:13 +0000 (0:00:00.668)       0:01:44.996 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-sb-db] *********************************************\nSaturday 21 March 2026  00:50:14 +0000 (0:00:00.991)       0:01:45.988 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Unset bootstrap args fact] **************************************\nSaturday 21 March 2026  00:50:15 +0000 (0:00:00.998)       0:01:46.986 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Ensuring config directories exist] ******************************\nSaturday 21 March 2026  00:50:16 +0000 (0:00:00.531)       0:01:47.518 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-db : Copying over config.json files for services] ********************\nSaturday 21 March 2026  00:50:17 +0000 (0:00:01.500)       0:01:49.018 ******** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-db : Check ovn containers] *******************************************\nSaturday 21 March 2026  00:50:21 +0000 (0:00:03.571)       0:01:52.590 ******** \n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-db : Flush handlers] *************************************************\nSaturday 21 March 2026  00:50:24 +0000 (0:00:02.940)       0:01:55.531 ******** \n\nTASK [ovn-db : Flush handlers] *************************************************\nSaturday 21 March 2026  00:50:24 +0000 (0:00:00.066)       0:01:55.597 ******** \n\nTASK [ovn-db : Flush handlers] *************************************************\nSaturday 21 March 2026  00:50:24 +0000 (0:00:00.065)       0:01:55.663 ******** \n\nRUNNING HANDLER [ovn-db : Restart ovn-nb-db container] *************************\nSaturday 21 March 2026  00:50:24 +0000 (0:00:00.362)       0:01:56.025 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ovn-db : Restart ovn-sb-db container] *************************\nSaturday 21 March 2026  00:50:30 +0000 (0:00:06.292)       0:02:02.318 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ovn-db : Restart ovn-northd container] ************************\nSaturday 21 March 2026  00:50:37 +0000 (0:00:06.190)       0:02:08.509 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ovn-db : Wait for leader election] ***************************************\nSaturday 21 March 2026  00:50:43 +0000 (0:00:06.125)       0:02:14.634 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Get OVN_Northbound cluster leader] ******************************\nSaturday 21 March 2026  00:50:43 +0000 (0:00:00.114)       0:02:14.749 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Configure OVN NB connection settings] ***************************\nSaturday 21 March 2026  00:50:44 +0000 (0:00:00.757)       0:02:15.507 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Get OVN_Southbound cluster leader] ******************************\nSaturday 21 March 2026  00:50:44 +0000 (0:00:00.703)       0:02:16.211 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Configure OVN SB connection settings] ***************************\nSaturday 21 March 2026  00:50:45 +0000 (0:00:00.771)       0:02:16.982 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-nb-db] *********************************************\nSaturday 21 March 2026  00:50:46 +0000 (0:00:00.705)       0:02:17.687 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-sb-db] *********************************************\nSaturday 21 March 2026  00:50:47 +0000 (0:00:00.840)       0:02:18.528 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=44  \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=43  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=22  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=43  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=22  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:50:48 +0000 (0:00:01.190)       0:02:19.719 ******** \n=============================================================================== \novn-controller : Restart ovn-controller container ---------------------- 27.41s\novn-controller : Configure OVN in OVSDB -------------------------------- 18.43s\novn-db : Restart ovn-sb-db container ----------------------------------- 13.89s\novn-db : Restart ovn-northd container ---------------------------------- 13.67s\novn-db : Restart ovn-nb-db container ------------------------------------ 9.01s\novn-db : Copying over config.json files for services -------------------- 4.18s\novn-db : Copying over config.json files for services -------------------- 3.57s\novn-db : Check ovn containers ------------------------------------------- 2.94s\novn-controller : Create br-int bridge on OpenvSwitch -------------------- 2.80s\novn-db : Check ovn containers ------------------------------------------- 2.41s\novn-controller : Check ovn-controller containers ------------------------ 2.03s\novn-controller : Ensuring systemd override directory exists ------------- 1.95s\novn-db : Get OVN_Northbound cluster leader ------------------------------ 1.95s\novn-controller : Copying over config.json files for services ------------ 1.75s\novn-controller : Reload systemd config ---------------------------------- 1.64s\novn-db : Ensuring config directories exist ------------------------------ 1.59s\novn-controller : Copying over systemd override -------------------------- 1.57s\novn-controller : Ensuring config directories exist ---------------------- 1.53s\novn-db : Ensuring config directories exist ------------------------------ 1.50s\nGroup hosts based on enabled services ----------------------------------- 1.36s\n\u001b[32m2026-03-21 00:50:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:50:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:50:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:50:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:50:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:51:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:51:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:51:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:51:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:51:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:51:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:51:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:51:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:51:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:51:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:51:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:51:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:51:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:51:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:51:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:51:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:51:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:51:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:51:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:51:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:52:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:52:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:52:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:52:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:52:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:52:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:52:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:52:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:52:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:52:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:52:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:52:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:52:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:52:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:52:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:52:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:52:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:52:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:52:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:52:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:52:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:53:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:53:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:53:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  00:47:19 +0000 (0:00:00.553)       0:00:00.553 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  00:47:19 +0000 (0:00:00.493)       0:00:01.046 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_loadbalancer_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_loadbalancer_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_loadbalancer_True)\u001b[0m\n\nPLAY [Apply role loadbalancer] *************************************************\n\nTASK [loadbalancer : include_tasks] ********************************************\nSaturday 21 March 2026  00:47:20 +0000 (0:00:00.481)       0:00:01.528 ******** \n\u001b[0;36mincluded: /ansible/roles/loadbalancer/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [loadbalancer : Check IPv6 support] ***************************************\nSaturday 21 March 2026  00:47:21 +0000 (0:00:01.501)       0:00:03.029 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Setting sysctl values] ***************************************************\nSaturday 21 March 2026  00:47:23 +0000 (0:00:01.593)       0:00:04.623 ******** \n\u001b[0;36mincluded: sysctl for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [sysctl : Check IPv6 support] *********************************************\nSaturday 21 March 2026  00:47:24 +0000 (0:00:00.707)       0:00:05.330 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [sysctl : Setting sysctl values] ******************************************\nSaturday 21 March 2026  00:47:25 +0000 (0:00:01.345)       0:00:06.676 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m\n\nTASK [module-load : Load modules] **********************************************\nSaturday 21 March 2026  00:47:28 +0000 (0:00:02.594)       0:00:09.271 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=ip_vs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=ip_vs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=ip_vs)\u001b[0m\n\nTASK [module-load : Persist modules via modules-load.d] ************************\nSaturday 21 March 2026  00:47:29 +0000 (0:00:00.969)       0:00:10.240 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=ip_vs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=ip_vs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=ip_vs)\u001b[0m\n\nTASK [module-load : Drop module persistence] ***********************************\nSaturday 21 March 2026  00:47:30 +0000 (0:00:01.394)       0:00:11.635 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=ip_vs) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=ip_vs) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=ip_vs) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Ensuring config directories exist] ************************\nSaturday 21 March 2026  00:47:31 +0000 (0:00:00.820)       0:00:12.456 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [loadbalancer : Ensuring haproxy service config subdir exists] ************\nSaturday 21 March 2026  00:47:32 +0000 (0:00:01.664)       0:00:14.120 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Ensuring proxysql service config subdirectories exist] ****\nSaturday 21 March 2026  00:47:34 +0000 (0:00:01.526)       0:00:15.647 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item=users)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=users)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=users)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=rules)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=rules)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=rules)\u001b[0m\n\nTASK [loadbalancer : Ensuring keepalived checks subdir exists] *****************\nSaturday 21 March 2026  00:47:36 +0000 (0:00:02.289)       0:00:17.937 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [loadbalancer : Remove mariadb.cfg if proxysql enabled] *******************\nSaturday 21 March 2026  00:47:38 +0000 (0:00:01.409)       0:00:19.346 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Removing checks for services which are disabled] **********\nSaturday 21 March 2026  00:47:40 +0000 (0:00:02.608)       0:00:21.955 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__dbae62d55c991b3958350c8dc51f9c592d9f63fa', '__omit_place_holder__dbae62d55c991b3958350c8dc51f9c592d9f63fa'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__dbae62d55c991b3958350c8dc51f9c592d9f63fa', '__omit_place_holder__dbae62d55c991b3958350c8dc51f9c592d9f63fa'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__dbae62d55c991b3958350c8dc51f9c592d9f63fa', '__omit_place_holder__dbae62d55c991b3958350c8dc51f9c592d9f63fa'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Copying checks for services which are enabled] ************\nSaturday 21 March 2026  00:47:43 +0000 (0:00:02.608)       0:00:24.564 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__dbae62d55c991b3958350c8dc51f9c592d9f63fa', '__omit_place_holder__dbae62d55c991b3958350c8dc51f9c592d9f63fa'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__dbae62d55c991b3958350c8dc51f9c592d9f63fa', '__omit_place_holder__dbae62d55c991b3958350c8dc51f9c592d9f63fa'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__dbae62d55c991b3958350c8dc51f9c592d9f63fa', '__omit_place_holder__dbae62d55c991b3958350c8dc51f9c592d9f63fa'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [loadbalancer : Copying over config.json files for services] **************\nSaturday 21 March 2026  00:47:46 +0000 (0:00:03.184)       0:00:27.748 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy.cfg] *********************************\nSaturday 21 March 2026  00:47:49 +0000 (0:00:02.868)       0:00:30.616 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m\n\nTASK [loadbalancer : Copying over proxysql config] *****************************\nSaturday 21 March 2026  00:47:51 +0000 (0:00:02.358)       0:00:32.975 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy single external frontend config] *****\nSaturday 21 March 2026  00:47:57 +0000 (0:00:05.487)       0:00:38.463 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Copying over custom haproxy services configuration] *******\nSaturday 21 March 2026  00:47:58 +0000 (0:00:00.939)       0:00:39.402 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m\n\nTASK [loadbalancer : Copying over keepalived.conf] *****************************\nSaturday 21 March 2026  00:48:00 +0000 (0:00:02.389)       0:00:41.792 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy.pem] *********************************\nSaturday 21 March 2026  00:48:02 +0000 (0:00:01.882)       0:00:43.675 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=haproxy.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=haproxy.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=haproxy.pem)\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy-internal.pem] ************************\nSaturday 21 March 2026  00:48:04 +0000 (0:00:01.630)       0:00:45.305 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=haproxy-internal.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=haproxy-internal.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=haproxy-internal.pem)\u001b[0m\n\nTASK [loadbalancer : include_tasks] ********************************************\nSaturday 21 March 2026  00:48:05 +0000 (0:00:01.550)       0:00:46.856 ******** \n\u001b[0;36mincluded: /ansible/roles/loadbalancer/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : loadbalancer | Copying over extra CA certificates] ***\nSaturday 21 March 2026  00:48:06 +0000 (0:00:00.834)       0:00:47.690 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-cert-copy : loadbalancer | Copying over backend internal TLS certificate] ***\nSaturday 21 March 2026  00:48:09 +0000 (0:00:03.340)       0:00:51.031 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : loadbalancer | Copying over backend internal TLS key] ***\nSaturday 21 March 2026  00:48:10 +0000 (0:00:00.531)       0:00:51.563 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over extra CA certificates] ********\nSaturday 21 March 2026  00:48:11 +0000 (0:00:01.446)       0:00:53.010 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] ***\nSaturday 21 March 2026  00:48:12 +0000 (0:00:00.653)       0:00:53.663 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over backend internal TLS key] *****\nSaturday 21 March 2026  00:48:13 +0000 (0:00:00.593)       0:00:54.257 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : proxysql | Copying over extra CA certificates] *******\nSaturday 21 March 2026  00:48:14 +0000 (0:00:01.238)       0:00:55.496 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : proxysql | Copying over backend internal TLS certificate] ***\nSaturday 21 March 2026  00:48:14 +0000 (0:00:00.637)       0:00:56.134 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : proxysql | Copying over backend internal TLS key] ****\nSaturday 21 March 2026  00:48:15 +0000 (0:00:00.599)       0:00:56.734 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy start script] ************************\nSaturday 21 March 2026  00:48:16 +0000 (0:00:01.387)       0:00:58.122 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m\n\nTASK [loadbalancer : Copying over proxysql start script] ***********************\nSaturday 21 March 2026  00:48:19 +0000 (0:00:02.494)       0:01:00.617 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m\n\nTASK [loadbalancer : Copying files for haproxy-ssh] ****************************\nSaturday 21 March 2026  00:48:20 +0000 (0:00:01.387)       0:01:02.004 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Check loadbalancer containers] ****************************\nSaturday 21 March 2026  00:48:21 +0000 (0:00:00.861)       0:01:02.866 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [include_role : aodh] *****************************************************\nSaturday 21 March 2026  00:48:23 +0000 (0:00:02.229)       0:01:05.095 ******** \n\u001b[0;36mincluded: aodh for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over aodh haproxy config] ***********************\nSaturday 21 March 2026  00:48:24 +0000 (0:00:00.566)       0:01:05.662 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for aodh when using single external frontend] ***\nSaturday 21 March 2026  00:48:27 +0000 (0:00:03.420)       0:01:09.082 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for aodh] **************************\nSaturday 21 March 2026  00:48:28 +0000 (0:00:00.714)       0:01:09.797 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [proxysql-config : Copying over aodh ProxySQL users config] ***************\nSaturday 21 March 2026  00:48:29 +0000 (0:00:00.802)       0:01:10.600 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over aodh ProxySQL rules config] ***************\nSaturday 21 March 2026  00:48:31 +0000 (0:00:01.829)       0:01:12.430 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : barbican] *************************************************\nSaturday 21 March 2026  00:48:33 +0000 (0:00:02.207)       0:01:14.637 ******** \n\u001b[0;36mincluded: barbican for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over barbican haproxy config] *******************\nSaturday 21 March 2026  00:48:34 +0000 (0:00:00.675)       0:01:15.313 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for barbican when using single external frontend] ***\nSaturday 21 March 2026  00:48:38 +0000 (0:00:04.283)       0:01:19.597 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for barbican] **********************\nSaturday 21 March 2026  00:48:39 +0000 (0:00:00.832)       0:01:20.429 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over barbican ProxySQL users config] ***********\nSaturday 21 March 2026  00:48:40 +0000 (0:00:00.990)       0:01:21.419 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over barbican ProxySQL rules config] ***********\nSaturday 21 March 2026  00:48:41 +0000 (0:00:01.311)       0:01:22.731 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : blazar] ***************************************************\nSaturday 21 March 2026  00:48:43 +0000 (0:00:01.975)       0:01:24.707 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : ceph-rgw] *************************************************\nSaturday 21 March 2026  00:48:43 +0000 (0:00:00.251)       0:01:24.959 ******** \n\u001b[0;36mincluded: ceph-rgw for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over ceph-rgw haproxy config] *******************\nSaturday 21 March 2026  00:48:46 +0000 (0:00:02.307)       0:01:27.267 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for ceph-rgw when using single external frontend] ***\nSaturday 21 March 2026  00:48:49 +0000 (0:00:02.946)       0:01:30.213 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for ceph-rgw] **********************\nSaturday 21 March 2026  00:48:50 +0000 (0:00:01.939)       0:01:32.153 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [proxysql-config : Copying over ceph-rgw ProxySQL users config] ***********\nSaturday 21 March 2026  00:48:53 +0000 (0:00:02.249)       0:01:34.402 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over ceph-rgw ProxySQL rules config] ***********\nSaturday 21 March 2026  00:48:53 +0000 (0:00:00.430)       0:01:34.832 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : cinder] ***************************************************\nSaturday 21 March 2026  00:48:54 +0000 (0:00:00.958)       0:01:35.791 ******** \n\u001b[0;36mincluded: cinder for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over cinder haproxy config] *********************\nSaturday 21 March 2026  00:48:55 +0000 (0:00:00.981)       0:01:36.773 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for cinder when using single external frontend] ***\nSaturday 21 March 2026  00:48:59 +0000 (0:00:03.753)       0:01:40.526 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for cinder] ************************\nSaturday 21 March 2026  00:48:59 +0000 (0:00:00.621)       0:01:41.148 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over cinder ProxySQL users config] *************\nSaturday 21 March 2026  00:49:00 +0000 (0:00:01.009)       0:01:42.157 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over cinder ProxySQL rules config] *************\nSaturday 21 March 2026  00:49:02 +0000 (0:00:01.053)       0:01:43.211 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [include_role : cloudkitty] ***********************************************\nSaturday 21 March 2026  00:49:03 +0000 (0:00:01.647)       0:01:44.858 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : cyborg] ***************************************************\nSaturday 21 March 2026  00:49:03 +0000 (0:00:00.238)       0:01:45.097 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : designate] ************************************************\nSaturday 21 March 2026  00:49:04 +0000 (0:00:00.257)       0:01:45.354 ******** \n\u001b[0;36mincluded: designate for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over designate haproxy config] ******************\nSaturday 21 March 2026  00:49:05 +0000 (0:00:00.915)       0:01:46.270 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for designate when using single external frontend] ***\nSaturday 21 March 2026  00:49:10 +0000 (0:00:05.051)       0:01:51.321 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for designate] *********************\nSaturday 21 March 2026  00:49:11 +0000 (0:00:01.464)       0:01:52.786 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over designate ProxySQL users config] **********\nSaturday 21 March 2026  00:49:13 +0000 (0:00:01.948)       0:01:54.734 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over designate ProxySQL rules config] **********\nSaturday 21 March 2026  00:49:14 +0000 (0:00:01.243)       0:01:55.977 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : etcd] *****************************************************\nSaturday 21 March 2026  00:49:16 +0000 (0:00:02.132)       0:01:58.109 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : glance] ***************************************************\nSaturday 21 March 2026  00:49:17 +0000 (0:00:00.285)       0:01:58.395 ******** \n\u001b[0;36mincluded: glance for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over glance haproxy config] *********************\nSaturday 21 March 2026  00:49:18 +0000 (0:00:00.938)       0:01:59.333 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for glance when using single external frontend] ***\nSaturday 21 March 2026  00:49:21 +0000 (0:00:03.828)       0:02:03.162 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for glance] ************************\nSaturday 21 March 2026  00:49:24 +0000 (0:00:02.525)       0:02:05.687 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over glance ProxySQL users config] *************\nSaturday 21 March 2026  00:49:27 +0000 (0:00:02.900)       0:02:08.587 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over glance ProxySQL rules config] *************\nSaturday 21 March 2026  00:49:28 +0000 (0:00:01.150)       0:02:09.738 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : gnocchi] **************************************************\nSaturday 21 March 2026  00:49:30 +0000 (0:00:01.665)       0:02:11.404 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : grafana] **************************************************\nSaturday 21 March 2026  00:49:30 +0000 (0:00:00.258)       0:02:11.662 ******** \n\u001b[0;36mincluded: grafana for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over grafana haproxy config] ********************\nSaturday 21 March 2026  00:49:31 +0000 (0:00:00.867)       0:02:12.530 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for grafana when using single external frontend] ***\nSaturday 21 March 2026  00:49:34 +0000 (0:00:02.816)       0:02:15.347 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for grafana] ***********************\nSaturday 21 March 2026  00:49:34 +0000 (0:00:00.398)       0:02:15.745 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over grafana ProxySQL users config] ************\nSaturday 21 March 2026  00:49:35 +0000 (0:00:00.791)       0:02:16.536 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over grafana ProxySQL rules config] ************\nSaturday 21 March 2026  00:49:36 +0000 (0:00:01.371)       0:02:17.908 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : heat] *****************************************************\nSaturday 21 March 2026  00:49:38 +0000 (0:00:02.001)       0:02:19.910 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : horizon] **************************************************\nSaturday 21 March 2026  00:49:38 +0000 (0:00:00.250)       0:02:20.160 ******** \n\u001b[0;36mincluded: horizon for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over horizon haproxy config] ********************\nSaturday 21 March 2026  00:49:39 +0000 (0:00:00.944)       0:02:21.104 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for horizon when using single external frontend] ***\nSaturday 21 March 2026  00:49:43 +0000 (0:00:03.190)       0:02:24.295 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for horizon] ***********************\nSaturday 21 March 2026  00:49:43 +0000 (0:00:00.589)       0:02:24.884 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over horizon ProxySQL users config] ************\nSaturday 21 March 2026  00:49:44 +0000 (0:00:00.860)       0:02:25.745 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over horizon ProxySQL rules config] ************\nSaturday 21 March 2026  00:49:46 +0000 (0:00:01.695)       0:02:27.440 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : influxdb] *************************************************\nSaturday 21 March 2026  00:49:48 +0000 (0:00:01.931)       0:02:29.372 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : ironic] ***************************************************\nSaturday 21 March 2026  00:49:48 +0000 (0:00:00.386)       0:02:29.758 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : keystone] *************************************************\nSaturday 21 March 2026  00:49:48 +0000 (0:00:00.272)       0:02:30.030 ******** \n\u001b[0;36mincluded: keystone for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over keystone haproxy config] *******************\nSaturday 21 March 2026  00:49:49 +0000 (0:00:01.006)       0:02:31.036 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for keystone when using single external frontend] ***\nSaturday 21 March 2026  00:49:53 +0000 (0:00:03.604)       0:02:34.641 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for keystone] **********************\nSaturday 21 March 2026  00:49:54 +0000 (0:00:00.629)       0:02:35.270 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over keystone ProxySQL users config] ***********\nSaturday 21 March 2026  00:49:55 +0000 (0:00:01.117)       0:02:36.387 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over keystone ProxySQL rules config] ***********\nSaturday 21 March 2026  00:49:56 +0000 (0:00:01.215)       0:02:37.603 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : letsencrypt] **********************************************\nSaturday 21 March 2026  00:49:58 +0000 (0:00:02.056)       0:02:39.660 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : magnum] ***************************************************\nSaturday 21 March 2026  00:49:58 +0000 (0:00:00.280)       0:02:39.941 ******** \n\u001b[0;36mincluded: magnum for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over magnum haproxy config] *********************\nSaturday 21 March 2026  00:49:59 +0000 (0:00:01.068)       0:02:41.009 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for magnum when using single external frontend] ***\nSaturday 21 March 2026  00:50:02 +0000 (0:00:03.124)       0:02:44.134 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for magnum] ************************\nSaturday 21 March 2026  00:50:03 +0000 (0:00:00.621)       0:02:44.756 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over magnum ProxySQL users config] *************\nSaturday 21 March 2026  00:50:04 +0000 (0:00:01.080)       0:02:45.836 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over magnum ProxySQL rules config] *************\nSaturday 21 March 2026  00:50:06 +0000 (0:00:01.358)       0:02:47.195 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : manila] ***************************************************\nSaturday 21 March 2026  00:50:08 +0000 (0:00:02.095)       0:02:49.290 ******** \n\u001b[0;36mincluded: manila for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over manila haproxy config] *********************\nSaturday 21 March 2026  00:50:09 +0000 (0:00:01.054)       0:02:50.345 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for manila when using single external frontend] ***\nSaturday 21 March 2026  00:50:13 +0000 (0:00:04.135)       0:02:54.481 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for manila] ************************\nSaturday 21 March 2026  00:50:14 +0000 (0:00:00.844)       0:02:55.325 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over manila ProxySQL users config] *************\nSaturday 21 March 2026  00:50:15 +0000 (0:00:01.076)       0:02:56.402 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over manila ProxySQL rules config] *************\nSaturday 21 March 2026  00:50:16 +0000 (0:00:01.431)       0:02:57.833 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : mariadb] **************************************************\nSaturday 21 March 2026  00:50:18 +0000 (0:00:02.182)       0:03:00.016 ******** \n\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [mariadb : Ensure mysql monitor user exist] *******************************\nSaturday 21 March 2026  00:50:20 +0000 (0:00:01.281)       0:03:01.298 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\nTASK [haproxy-config : Copying over mariadb haproxy config] ********************\nSaturday 21 March 2026  00:50:23 +0000 (0:00:03.128)       0:03:04.427 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [haproxy-config : Add configuration for mariadb when using single external frontend] ***\nSaturday 21 March 2026  00:50:25 +0000 (0:00:02.199)       0:03:06.627 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for mariadb] ***********************\nSaturday 21 March 2026  00:50:27 +0000 (0:00:02.163)       0:03:08.790 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over mariadb ProxySQL users config] ************\nSaturday 21 March 2026  00:50:29 +0000 (0:00:02.027)       0:03:10.817 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over mariadb ProxySQL rules config] ************\nSaturday 21 March 2026  00:50:31 +0000 (0:00:01.840)       0:03:12.658 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : masakari] *************************************************\nSaturday 21 March 2026  00:50:32 +0000 (0:00:01.397)       0:03:14.056 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : memcached] ************************************************\nSaturday 21 March 2026  00:50:33 +0000 (0:00:00.272)       0:03:14.328 ******** \n\u001b[0;36mincluded: memcached for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over memcached haproxy config] ******************\nSaturday 21 March 2026  00:50:34 +0000 (0:00:01.179)       0:03:15.508 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for memcached when using single external frontend] ***\nSaturday 21 March 2026  00:50:35 +0000 (0:00:01.427)       0:03:16.936 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for memcached] *********************\nSaturday 21 March 2026  00:50:36 +0000 (0:00:00.349)       0:03:17.285 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over memcached ProxySQL users config] **********\nSaturday 21 March 2026  00:50:36 +0000 (0:00:00.734)       0:03:18.020 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over memcached ProxySQL rules config] **********\nSaturday 21 March 2026  00:50:37 +0000 (0:00:00.432)       0:03:18.453 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : mistral] **************************************************\nSaturday 21 March 2026  00:50:38 +0000 (0:00:01.074)       0:03:19.527 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : neutron] **************************************************\nSaturday 21 March 2026  00:50:38 +0000 (0:00:00.264)       0:03:19.792 ******** \n\u001b[0;36mincluded: neutron for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over neutron haproxy config] ********************\nSaturday 21 March 2026  00:50:39 +0000 (0:00:01.273)       0:03:21.066 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for neutron when using single external frontend] ***\nSaturday 21 March 2026  00:50:43 +0000 (0:00:03.957)       0:03:25.023 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for neutron] ***********************\nSaturday 21 March 2026  00:50:45 +0000 (0:00:01.820)       0:03:26.844 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over neutron ProxySQL users config] ************\nSaturday 21 March 2026  00:50:47 +0000 (0:00:01.427)       0:03:28.271 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over neutron ProxySQL rules config] ************\nSaturday 21 March 2026  00:50:48 +0000 (0:00:01.339)       0:03:29.610 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : placement] ************************************************\nSaturday 21 March 2026  00:50:50 +0000 (0:00:01.921)       0:03:31.532 ******** \n\u001b[0;36mincluded: placement for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over placement haproxy config] ******************\nSaturday 21 March 2026  00:50:51 +0000 (0:00:01.231)       0:03:32.763 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for placement when using single external frontend] ***\nSaturday 21 March 2026  00:50:54 +0000 (0:00:02.896)       0:03:35.660 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for placement] *********************\nSaturday 21 March 2026  00:50:54 +0000 (0:00:00.428)       0:03:36.088 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over placement ProxySQL users config] **********\nSaturday 21 March 2026  00:50:55 +0000 (0:00:01.058)       0:03:37.146 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over placement ProxySQL rules config] **********\nSaturday 21 March 2026  00:50:57 +0000 (0:00:01.359)       0:03:38.506 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : nova] *****************************************************\nSaturday 21 March 2026  00:50:59 +0000 (0:00:01.931)       0:03:40.437 ******** \n\u001b[0;36mincluded: nova for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over nova haproxy config] ***********************\nSaturday 21 March 2026  00:51:00 +0000 (0:00:01.314)       0:03:41.752 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for nova when using single external frontend] ***\nSaturday 21 March 2026  00:51:04 +0000 (0:00:03.693)       0:03:45.446 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for nova] **************************\nSaturday 21 March 2026  00:51:04 +0000 (0:00:00.564)       0:03:46.010 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova ProxySQL users config] ***************\nSaturday 21 March 2026  00:51:05 +0000 (0:00:00.818)       0:03:46.829 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova ProxySQL rules config] ***************\nSaturday 21 March 2026  00:51:07 +0000 (0:00:01.647)       0:03:48.476 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : nova-cell] ************************************************\nSaturday 21 March 2026  00:51:09 +0000 (0:00:02.002)       0:03:50.479 ******** \n\u001b[0;36mincluded: nova-cell for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova-cell : Configure loadbalancer for nova-novncproxy] ******************\nSaturday 21 March 2026  00:51:10 +0000 (0:00:01.248)       0:03:51.727 ******** \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-novncproxy)\u001b[0m\n\nTASK [haproxy-config : Copying over nova-cell:nova-novncproxy haproxy config] ***\nSaturday 21 March 2026  00:51:11 +0000 (0:00:01.122)       0:03:52.850 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for nova-cell:nova-novncproxy when using single external frontend] ***\nSaturday 21 March 2026  00:51:15 +0000 (0:00:03.420)       0:03:56.271 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for nova-cell:nova-novncproxy] *****\nSaturday 21 March 2026  00:51:16 +0000 (0:00:01.502)       0:03:57.773 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********\nSaturday 21 March 2026  00:51:18 +0000 (0:00:01.880)       0:03:59.654 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********\nSaturday 21 March 2026  00:51:20 +0000 (0:00:02.244)       0:04:01.898 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Configure loadbalancer for nova-spicehtml5proxy] *************\nSaturday 21 March 2026  00:51:23 +0000 (0:00:02.605)       0:04:04.503 ******** \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-spicehtml5proxy)\u001b[0m\n\nTASK [haproxy-config : Copying over nova-cell:nova-spicehtml5proxy haproxy config] ***\nSaturday 21 March 2026  00:51:24 +0000 (0:00:00.759)       0:04:05.262 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Add configuration for nova-cell:nova-spicehtml5proxy when using single external frontend] ***\nSaturday 21 March 2026  00:51:25 +0000 (0:00:01.112)       0:04:06.375 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for nova-cell:nova-spicehtml5proxy] ***\nSaturday 21 March 2026  00:51:26 +0000 (0:00:01.327)       0:04:07.702 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********\nSaturday 21 March 2026  00:51:27 +0000 (0:00:01.299)       0:04:09.001 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********\nSaturday 21 March 2026  00:51:30 +0000 (0:00:02.259)       0:04:11.261 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [nova-cell : Configure loadbalancer for nova-serialproxy] *****************\nSaturday 21 March 2026  00:51:32 +0000 (0:00:02.861)       0:04:14.122 ******** \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-serialproxy)\u001b[0m\n\nTASK [haproxy-config : Copying over nova-cell:nova-serialproxy haproxy config] ***\nSaturday 21 March 2026  00:51:33 +0000 (0:00:00.849)       0:04:14.971 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Add configuration for nova-cell:nova-serialproxy when using single external frontend] ***\nSaturday 21 March 2026  00:51:34 +0000 (0:00:01.146)       0:04:16.118 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for nova-cell:nova-serialproxy] ****\nSaturday 21 March 2026  00:51:36 +0000 (0:00:01.116)       0:04:17.235 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********\nSaturday 21 March 2026  00:51:37 +0000 (0:00:01.287)       0:04:18.522 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********\nSaturday 21 March 2026  00:51:39 +0000 (0:00:02.458)       0:04:20.981 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [include_role : octavia] **************************************************\nSaturday 21 March 2026  00:51:42 +0000 (0:00:02.949)       0:04:23.930 ******** \n\u001b[0;36mincluded: octavia for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over octavia haproxy config] ********************\nSaturday 21 March 2026  00:51:43 +0000 (0:00:01.220)       0:04:25.151 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for octavia when using single external frontend] ***\nSaturday 21 March 2026  00:51:47 +0000 (0:00:03.471)       0:04:28.622 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for octavia] ***********************\nSaturday 21 March 2026  00:51:48 +0000 (0:00:00.877)       0:04:29.500 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over octavia ProxySQL users config] ************\nSaturday 21 March 2026  00:51:49 +0000 (0:00:00.920)       0:04:30.420 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over octavia ProxySQL rules config] ************\nSaturday 21 March 2026  00:51:50 +0000 (0:00:01.318)       0:04:31.739 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : opensearch] ***********************************************\nSaturday 21 March 2026  00:51:52 +0000 (0:00:02.090)       0:04:33.830 ******** \n\u001b[0;36mincluded: opensearch for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over opensearch haproxy config] *****************\nSaturday 21 March 2026  00:51:54 +0000 (0:00:01.597)       0:04:35.427 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for opensearch when using single external frontend] ***\nSaturday 21 March 2026  00:51:58 +0000 (0:00:04.726)       0:04:40.153 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for opensearch] ********************\nSaturday 21 March 2026  00:51:59 +0000 (0:00:00.809)       0:04:40.962 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over opensearch ProxySQL users config] *********\nSaturday 21 March 2026  00:52:00 +0000 (0:00:01.056)       0:04:42.019 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over opensearch ProxySQL rules config] *********\nSaturday 21 March 2026  00:52:01 +0000 (0:00:00.417)       0:04:42.437 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : prometheus] ***********************************************\nSaturday 21 March 2026  00:52:02 +0000 (0:00:01.126)       0:04:43.563 ******** \n\u001b[0;36mincluded: prometheus for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over prometheus haproxy config] *****************\nSaturday 21 March 2026  00:52:03 +0000 (0:00:01.541)       0:04:45.105 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for prometheus when using single external frontend] ***\nSaturday 21 March 2026  00:52:08 +0000 (0:00:04.081)       0:04:49.187 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for prometheus] ********************\nSaturday 21 March 2026  00:52:08 +0000 (0:00:00.761)       0:04:49.948 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over prometheus ProxySQL users config] *********\nSaturday 21 March 2026  00:52:09 +0000 (0:00:01.047)       0:04:50.996 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over prometheus ProxySQL rules config] *********\nSaturday 21 March 2026  00:52:10 +0000 (0:00:00.428)       0:04:51.424 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : rabbitmq] *************************************************\nSaturday 21 March 2026  00:52:11 +0000 (0:00:01.111)       0:04:52.536 ******** \n\u001b[0;36mincluded: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over rabbitmq haproxy config] *******************\nSaturday 21 March 2026  00:52:12 +0000 (0:00:01.322)       0:04:53.858 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for rabbitmq when using single external frontend] ***\nSaturday 21 March 2026  00:52:14 +0000 (0:00:02.105)       0:04:55.964 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for rabbitmq] **********************\nSaturday 21 March 2026  00:52:15 +0000 (0:00:00.395)       0:04:56.359 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over rabbitmq ProxySQL users config] ***********\nSaturday 21 March 2026  00:52:15 +0000 (0:00:00.577)       0:04:56.937 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over rabbitmq ProxySQL rules config] ***********\nSaturday 21 March 2026  00:52:16 +0000 (0:00:00.656)       0:04:57.593 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : skyline] **************************************************\nSaturday 21 March 2026  00:52:17 +0000 (0:00:01.117)       0:04:58.711 ******** \n\u001b[0;36mincluded: skyline for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over skyline haproxy config] ********************\nSaturday 21 March 2026  00:52:18 +0000 (0:00:01.405)       0:05:00.116 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for skyline when using single external frontend] ***\nSaturday 21 March 2026  00:52:24 +0000 (0:00:05.240)       0:05:05.357 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for skyline] ***********************\nSaturday 21 March 2026  00:52:25 +0000 (0:00:00.830)       0:05:06.188 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over skyline ProxySQL users config] ************\nSaturday 21 March 2026  00:52:25 +0000 (0:00:00.833)       0:05:07.021 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over skyline ProxySQL rules config] ************\nSaturday 21 March 2026  00:52:27 +0000 (0:00:01.330)       0:05:08.352 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : swift] ****************************************************\nSaturday 21 March 2026  00:52:29 +0000 (0:00:02.079)       0:05:10.432 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : tacker] ***************************************************\nSaturday 21 March 2026  00:52:29 +0000 (0:00:00.477)       0:05:10.910 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : trove] ****************************************************\nSaturday 21 March 2026  00:52:29 +0000 (0:00:00.268)       0:05:11.178 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : venus] ****************************************************\nSaturday 21 March 2026  00:52:30 +0000 (0:00:00.282)       0:05:11.461 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : watcher] **************************************************\nSaturday 21 March 2026  00:52:30 +0000 (0:00:00.264)       0:05:11.726 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : zun] ******************************************************\nSaturday 21 March 2026  00:52:31 +0000 (0:00:00.467)       0:05:12.193 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Check IP addresses on the API interface] *******\nSaturday 21 March 2026  00:52:31 +0000 (0:00:00.491)       0:05:12.685 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Group HA nodes by status] **********************\nSaturday 21 March 2026  00:52:32 +0000 (0:00:00.663)       0:05:13.348 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop backup keepalived container] **************\nSaturday 21 March 2026  00:52:32 +0000 (0:00:00.496)       0:05:13.845 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop backup haproxy container] *****************\nSaturday 21 March 2026  00:52:33 +0000 (0:00:00.901)       0:05:14.746 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop backup proxysql container] ****************\nSaturday 21 March 2026  00:52:34 +0000 (0:00:01.030)       0:05:15.777 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start backup haproxy container] ****************\nSaturday 21 March 2026  00:52:35 +0000 (0:00:00.969)       0:05:16.746 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for backup haproxy to start] **************\nSaturday 21 March 2026  00:52:39 +0000 (0:00:04.268)       0:05:21.015 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start backup proxysql container] ***************\nSaturday 21 March 2026  00:52:42 +0000 (0:00:03.032)       0:05:24.047 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for backup proxysql to start] *************\nSaturday 21 March 2026  00:52:51 +0000 (0:00:08.295)       0:05:32.343 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start backup keepalived container] *************\nSaturday 21 March 2026  00:52:54 +0000 (0:00:03.751)       0:05:36.094 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop master haproxy container] *****************\nSaturday 21 March 2026  00:52:59 +0000 (0:00:04.599)       0:05:40.694 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop master proxysql container] ****************\nSaturday 21 March 2026  00:53:00 +0000 (0:00:00.534)       0:05:41.228 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop master keepalived container] **************\nSaturday 21 March 2026  00:53:00 +0000 (0:00:00.312)       0:05:41.541 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start master haproxy container] ****************\nSaturday 21 March 2026  00:53:00 +0000 (0:00:00.335)       0:05:41.876 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start master proxysql container] ***************\nSaturday 21 March 2026  00:53:01 +0000 (0:00:00.315)       0:05:42.192 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start master keepalived container] *************\nSaturday 21 March 2026  00:53:01 +0000 (0:00:00.546)       0:05:42.739 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for haproxy to listen on VIP] *************\nSaturday 21 March 2026  00:53:01 +0000 (0:00:00.302)       0:05:43.041 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for proxysql to listen on VIP] ************\nSaturday 21 March 2026  00:53:06 +0000 (0:00:04.728)       0:05:47.770 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=123 \u001b[0m \u001b[0;33mchanged=76  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=97  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=122 \u001b[0m \u001b[0;33mchanged=76  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=97  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=122 \u001b[0m \u001b[0;33mchanged=76  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=97  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:53:07 +0000 (0:00:00.868)       0:05:48.639 ******** \n=============================================================================== \nloadbalancer : Start backup proxysql container -------------------------- 8.30s\nloadbalancer : Copying over proxysql config ----------------------------- 5.49s\nhaproxy-config : Copying over skyline haproxy config -------------------- 5.24s\nhaproxy-config : Copying over designate haproxy config ------------------ 5.05s\nloadbalancer : Wait for haproxy to listen on VIP ------------------------ 4.73s\nhaproxy-config : Copying over opensearch haproxy config ----------------- 4.73s\nloadbalancer : Start backup keepalived container ------------------------ 4.60s\nhaproxy-config : Copying over barbican haproxy config ------------------- 4.28s\nloadbalancer : Start backup haproxy container --------------------------- 4.27s\nhaproxy-config : Copying over manila haproxy config --------------------- 4.14s\nhaproxy-config : Copying over prometheus haproxy config ----------------- 4.08s\nhaproxy-config : Copying over neutron haproxy config -------------------- 3.96s\nhaproxy-config : Copying over glance haproxy config --------------------- 3.83s\nhaproxy-config : Copying over cinder haproxy config --------------------- 3.75s\nloadbalancer : Wait for backup proxysql to start ------------------------ 3.75s\nhaproxy-config : Copying over nova haproxy config ----------------------- 3.69s\nhaproxy-config : Copying over keystone haproxy config ------------------- 3.60s\nhaproxy-config : Copying over octavia haproxy config -------------------- 3.47s\nhaproxy-config : Copying over nova-cell:nova-novncproxy haproxy config --- 3.42s\nhaproxy-config : Copying over aodh haproxy config ----------------------- 3.42s\n\u001b[32m2026-03-21 00:53:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:53:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:53:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:53:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:53:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:53:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:53:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:53:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:53:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:53:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:53:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:53:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:53:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:53:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:53:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:53:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:53:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:53:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:54:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:54:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:54:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:54:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:54:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:54:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:54:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:54:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:54:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:54:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:54:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:54:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:54:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:54:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:54:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:54:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:54:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:54:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:54:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:54:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:54:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:55:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:55:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:55:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:55:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:55:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:55:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:55:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:55:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:55:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:55:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:55:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:55:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:55:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:55:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:55:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:55:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-21 00:55:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  00:53:11 +0000 (0:00:00.284)       0:00:00.284 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  00:53:11 +0000 (0:00:00.269)       0:00:00.553 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_opensearch_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_opensearch_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_opensearch_True)\u001b[0m\n\nPLAY [Apply role opensearch] ***************************************************\n\nTASK [opensearch : include_tasks] **********************************************\nSaturday 21 March 2026  00:53:11 +0000 (0:00:00.279)       0:00:00.833 ******** \n\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [opensearch : Setting sysctl values] **************************************\nSaturday 21 March 2026  00:53:12 +0000 (0:00:00.583)       0:00:01.416 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\nTASK [opensearch : Ensuring config directories exist] **************************\nSaturday 21 March 2026  00:53:13 +0000 (0:00:00.937)       0:00:02.354 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\nTASK [opensearch : include_tasks] **********************************************\nSaturday 21 March 2026  00:53:14 +0000 (0:00:01.284)       0:00:03.638 ******** \n\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : opensearch | Copying over extra CA certificates] *****\nSaturday 21 March 2026  00:53:15 +0000 (0:00:00.585)       0:00:04.224 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\nTASK [service-cert-copy : opensearch | Copying over backend internal TLS certificate] ***\nSaturday 21 March 2026  00:53:18 +0000 (0:00:03.593)       0:00:07.817 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : opensearch | Copying over backend internal TLS key] ***\nSaturday 21 March 2026  00:53:19 +0000 (0:00:00.777)       0:00:08.595 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [opensearch : Copying over config.json files for services] ****************\nSaturday 21 March 2026  00:53:20 +0000 (0:00:00.769)       0:00:09.364 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\nTASK [opensearch : Copying over opensearch service config file] ****************\nSaturday 21 March 2026  00:53:22 +0000 (0:00:02.721)       0:00:12.085 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [opensearch : Copying over opensearch-dashboards config file] *************\nSaturday 21 March 2026  00:53:25 +0000 (0:00:02.984)       0:00:15.069 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [opensearch : Check opensearch containers] ********************************\nSaturday 21 March 2026  00:53:27 +0000 (0:00:01.500)       0:00:16.570 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\nTASK [opensearch : include_tasks] **********************************************\nSaturday 21 March 2026  00:53:29 +0000 (0:00:02.237)       0:00:18.808 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [opensearch : Flush handlers] *********************************************\nSaturday 21 March 2026  00:53:30 +0000 (0:00:00.427)       0:00:19.235 ******** \n\nTASK [opensearch : Flush handlers] *********************************************\nSaturday 21 March 2026  00:53:30 +0000 (0:00:00.066)       0:00:19.302 ******** \n\nTASK [opensearch : Flush handlers] *********************************************\nSaturday 21 March 2026  00:53:30 +0000 (0:00:00.065)       0:00:19.367 ******** \n\nRUNNING HANDLER [opensearch : Disable shard allocation] ************************\nSaturday 21 March 2026  00:53:30 +0000 (0:00:00.067)       0:00:19.435 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [opensearch : Perform a flush] *********************************\nSaturday 21 March 2026  00:53:30 +0000 (0:00:00.243)       0:00:19.678 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [opensearch : Restart opensearch container] ********************\nSaturday 21 March 2026  00:53:30 +0000 (0:00:00.237)       0:00:19.916 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [opensearch : Restart opensearch-dashboards container] *********\nSaturday 21 March 2026  00:54:27 +0000 (0:00:56.884)       0:01:16.801 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [opensearch : include_tasks] **********************************************\nSaturday 21 March 2026  00:55:34 +0000 (0:01:07.297)       0:02:24.098 ******** \n\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/post-config.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [opensearch : Wait for OpenSearch to become ready] ************************\nSaturday 21 March 2026  00:55:35 +0000 (0:00:00.666)       0:02:24.765 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [opensearch : Wait for OpenSearch cluster to become healthy] **************\nSaturday 21 March 2026  00:55:37 +0000 (0:00:02.292)       0:02:27.057 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [opensearch : Check if a log retention policy exists] *********************\nSaturday 21 March 2026  00:55:39 +0000 (0:00:01.932)       0:02:28.989 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [opensearch : Create new log retention policy] ****************************\nSaturday 21 March 2026  00:55:41 +0000 (0:00:01.973)       0:02:30.963 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [opensearch : Apply retention policy to existing indices] *****************\nSaturday 21 March 2026  00:55:44 +0000 (0:00:02.273)       0:02:33.237 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:55:46 +0000 (0:00:02.166)       0:02:35.403 ******** \n=============================================================================== \nopensearch : Restart opensearch-dashboards container ------------------- 67.30s\nopensearch : Restart opensearch container ------------------------------ 56.88s\nservice-cert-copy : opensearch | Copying over extra CA certificates ----- 3.59s\nopensearch : Copying over opensearch service config file ---------------- 2.98s\nopensearch : Copying over config.json files for services ---------------- 2.72s\nopensearch : Wait for OpenSearch to become ready ------------------------ 2.29s\nopensearch : Create new log retention policy ---------------------------- 2.27s\nopensearch : Check opensearch containers -------------------------------- 2.24s\nopensearch : Apply retention policy to existing indices ----------------- 2.17s\nopensearch : Check if a log retention policy exists --------------------- 1.97s\nopensearch : Wait for OpenSearch cluster to become healthy -------------- 1.93s\nopensearch : Copying over opensearch-dashboards config file ------------- 1.50s\nopensearch : Ensuring config directories exist -------------------------- 1.28s\nopensearch : Setting sysctl values -------------------------------------- 0.94s\nservice-cert-copy : opensearch | Copying over backend internal TLS certificate --- 0.78s\nservice-cert-copy : opensearch | Copying over backend internal TLS key --- 0.77s\nopensearch : include_tasks ---------------------------------------------- 0.67s\nopensearch : include_tasks ---------------------------------------------- 0.59s\nopensearch : include_tasks ---------------------------------------------- 0.58s\nopensearch : include_tasks ---------------------------------------------- 0.43s\n\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Prepare deployment of Ceph services] *************************************\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nSaturday 21 March 2026  00:44:39 +0000 (0:00:00.774)       0:00:00.774 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nSaturday 21 March 2026  00:44:40 +0000 (0:00:01.331)       0:00:02.106 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nSaturday 21 March 2026  00:44:42 +0000 (0:00:02.059)       0:00:04.165 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nSaturday 21 March 2026  00:44:43 +0000 (0:00:00.840)       0:00:05.006 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nSaturday 21 March 2026  00:44:45 +0000 (0:00:01.496)       0:00:06.502 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nSaturday 21 March 2026  00:44:46 +0000 (0:00:01.230)       0:00:07.733 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nSaturday 21 March 2026  00:44:47 +0000 (0:00:00.700)       0:00:08.433 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nSaturday 21 March 2026  00:44:47 +0000 (0:00:00.921)       0:00:09.355 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nSaturday 21 March 2026  00:44:48 +0000 (0:00:00.926)       0:00:10.282 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nSaturday 21 March 2026  00:44:50 +0000 (0:00:01.173)       0:00:11.455 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nSaturday 21 March 2026  00:44:50 +0000 (0:00:00.743)       0:00:12.198 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nSaturday 21 March 2026  00:44:51 +0000 (0:00:01.166)       0:00:13.365 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nSaturday 21 March 2026  00:44:54 +0000 (0:00:02.872)       0:00:16.237 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nSaturday 21 March 2026  00:44:55 +0000 (0:00:00.627)       0:00:16.864 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nSaturday 21 March 2026  00:44:57 +0000 (0:00:01.603)       0:00:18.467 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nSaturday 21 March 2026  00:44:57 +0000 (0:00:00.246)       0:00:18.714 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-21 00:44:52.964412', 'end': '2026-03-21 00:44:53.064620', 'delta': '0:00:00.100208', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-21 00:44:53.564597', 'end': '2026-03-21 00:44:53.659889', 'delta': '0:00:00.095292', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-21 00:44:54.311062', 'end': '2026-03-21 00:44:54.406291', 'delta': '0:00:00.095229', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nSaturday 21 March 2026  00:44:57 +0000 (0:00:00.693)       0:00:19.407 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nSaturday 21 March 2026  00:45:00 +0000 (0:00:02.996)       0:00:22.404 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nSaturday 21 March 2026  00:45:01 +0000 (0:00:00.878)       0:00:23.282 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nSaturday 21 March 2026  00:45:03 +0000 (0:00:01.636)       0:00:24.919 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nSaturday 21 March 2026  00:45:04 +0000 (0:00:01.023)       0:00:25.943 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nSaturday 21 March 2026  00:45:05 +0000 (0:00:01.172)       0:00:27.115 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nSaturday 21 March 2026  00:45:05 +0000 (0:00:00.285)       0:00:27.400 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nSaturday 21 March 2026  00:45:06 +0000 (0:00:00.448)       0:00:27.848 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nSaturday 21 March 2026  00:45:07 +0000 (0:00:01.061)       0:00:28.910 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nSaturday 21 March 2026  00:45:08 +0000 (0:00:00.754)       0:00:29.664 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nSaturday 21 March 2026  00:45:09 +0000 (0:00:00.871)       0:00:30.535 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nSaturday 21 March 2026  00:45:09 +0000 (0:00:00.831)       0:00:31.367 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nSaturday 21 March 2026  00:45:10 +0000 (0:00:00.713)       0:00:32.080 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nSaturday 21 March 2026  00:45:12 +0000 (0:00:01.479)       0:00:33.559 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nSaturday 21 March 2026  00:45:12 +0000 (0:00:00.656)       0:00:34.216 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--64dd100a--6bcc--5763--b557--0732c79bfd55-osd--block--64dd100a--6bcc--5763--b557--0732c79bfd55', 'dm-uuid-LVM-MWn87OlXxZ0hjMuGG89puNwiFbMviFcm1xAms0qPz8LYU8Ut0AGzeF9WoI61D5NM'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--5368e13a--d21f--5912--91bd--da537815d9e8-osd--block--5368e13a--d21f--5912--91bd--da537815d9e8', 'dm-uuid-LVM-xgLPNxLNUSwUnkWnyYv6gd71t6QIH1sH9rPUXF3date4d4aBKBICcuqoW7mNU6TI'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part1', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part14', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part15', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part16', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'holders': ['ceph--64dd100a--6bcc--5763--b557--0732c79bfd55-osd--block--64dd100a--6bcc--5763--b557--0732c79bfd55'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-HiPrG3-9qzo-WJZ0-1Smx-Bi1t-ReBz-6cdxta', 'scsi-0QEMU_QEMU_HARDDISK_f08311eb-e2ad-4eb7-8da9-ac6d899a2409', 'scsi-SQEMU_QEMU_HARDDISK_f08311eb-e2ad-4eb7-8da9-ac6d899a2409'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'holders': ['ceph--5368e13a--d21f--5912--91bd--da537815d9e8-osd--block--5368e13a--d21f--5912--91bd--da537815d9e8'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-m0F2nF-j9UB-mZfT-bRv8-xW1c-EVch-ZfDvHt', 'scsi-0QEMU_QEMU_HARDDISK_e03b171b-b829-41a0-a0b6-cc46d85f013d', 'scsi-SQEMU_QEMU_HARDDISK_e03b171b-b829-41a0-a0b6-cc46d85f013d'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_9fa3c153-0e2a-45ef-8e48-e0095eb937ad', 'scsi-SQEMU_QEMU_HARDDISK_9fa3c153-0e2a-45ef-8e48-e0095eb937ad'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-21-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--ed04aa0b--4f5a--523b--a5d2--05072a1b260d-osd--block--ed04aa0b--4f5a--523b--a5d2--05072a1b260d', 'dm-uuid-LVM-viMZkgfP3jZUgTAo0bSRpd40fLtFkJJmO8XW5mmi1aRChVrq8X3DDkwGl8XRb5at'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--45b09425--beda--50b4--a83b--1fc50b5adca4-osd--block--45b09425--beda--50b4--a83b--1fc50b5adca4', 'dm-uuid-LVM-XzfTGfmLHoftdDBti3ot8l7sbRObO8yALj7bfZRysK7QQ9WjzIgwF2E0nO1Pc63W'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part1', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part14', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part15', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part16', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'holders': ['ceph--ed04aa0b--4f5a--523b--a5d2--05072a1b260d-osd--block--ed04aa0b--4f5a--523b--a5d2--05072a1b260d'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-nra7ru-O0w6-dM26-lz5o-2arZ-rXsE-lANagX', 'scsi-0QEMU_QEMU_HARDDISK_4e74a544-8b3a-4a05-98a1-b7cc13f13c18', 'scsi-SQEMU_QEMU_HARDDISK_4e74a544-8b3a-4a05-98a1-b7cc13f13c18'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'holders': ['ceph--45b09425--beda--50b4--a83b--1fc50b5adca4-osd--block--45b09425--beda--50b4--a83b--1fc50b5adca4'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-XQTe4H-QTjK-DLSs-Qjj6-RKjC-da4T-xXChra', 'scsi-0QEMU_QEMU_HARDDISK_666e73fe-87f5-4f65-9ff2-95a038c8b18d', 'scsi-SQEMU_QEMU_HARDDISK_666e73fe-87f5-4f65-9ff2-95a038c8b18d'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ae4aabb-0174-4999-a1d0-4bf78cbb2965', 'scsi-SQEMU_QEMU_HARDDISK_8ae4aabb-0174-4999-a1d0-4bf78cbb2965'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-14-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--ede169e1--d974--55ed--a938--c303cb89cb5b-osd--block--ede169e1--d974--55ed--a938--c303cb89cb5b', 'dm-uuid-LVM-bEBXWAqtQdfT6xKYbL0dF10We9CLdKgKh0pf0vpR2lQnL405kp4ndXs76m3o0J5Y'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a-osd--block--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a', 'dm-uuid-LVM-G1LwYmzPCvwLZRWX1y5oYSHlzZdIxlOhfbzzvktDUwfRIgHaiaw6r0MV8HVh1TLk'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part1', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part14', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part15', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part16', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'holders': ['ceph--ede169e1--d974--55ed--a938--c303cb89cb5b-osd--block--ede169e1--d974--55ed--a938--c303cb89cb5b'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-HVhEyD-n8kM-PvHm-CUDD-tuRZ-igjC-Nzour2', 'scsi-0QEMU_QEMU_HARDDISK_f76ae8f0-65cc-4a62-b19c-23f8a40fbde6', 'scsi-SQEMU_QEMU_HARDDISK_f76ae8f0-65cc-4a62-b19c-23f8a40fbde6'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'holders': ['ceph--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a-osd--block--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-YifJdB-JYWE-HhRV-mF1g-0k83-hzWG-ctJZWy', 'scsi-0QEMU_QEMU_HARDDISK_d7bb81ba-6491-4b5a-bb82-f0b49314eac0', 'scsi-SQEMU_QEMU_HARDDISK_d7bb81ba-6491-4b5a-bb82-f0b49314eac0'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_cdffe5f1-7f0d-460e-a1db-d6846d226144', 'scsi-SQEMU_QEMU_HARDDISK_cdffe5f1-7f0d-460e-a1db-d6846d226144'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-28-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f', 'scsi-SQEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part1', 'scsi-SQEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part14', 'scsi-SQEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part15', 'scsi-SQEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part16', 'scsi-SQEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-10-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5', 'scsi-SQEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part1', 'scsi-SQEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part14', 'scsi-SQEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part15', 'scsi-SQEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part16', 'scsi-SQEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-23-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32', 'scsi-SQEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part1', 'scsi-SQEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part14', 'scsi-SQEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part15', 'scsi-SQEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part16', 'scsi-SQEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-19-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nSaturday 21 March 2026  00:45:14 +0000 (0:00:01.751)       0:00:35.968 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--64dd100a--6bcc--5763--b557--0732c79bfd55-osd--block--64dd100a--6bcc--5763--b557--0732c79bfd55', 'dm-uuid-LVM-MWn87OlXxZ0hjMuGG89puNwiFbMviFcm1xAms0qPz8LYU8Ut0AGzeF9WoI61D5NM'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--5368e13a--d21f--5912--91bd--da537815d9e8-osd--block--5368e13a--d21f--5912--91bd--da537815d9e8', 'dm-uuid-LVM-xgLPNxLNUSwUnkWnyYv6gd71t6QIH1sH9rPUXF3date4d4aBKBICcuqoW7mNU6TI'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--ed04aa0b--4f5a--523b--a5d2--05072a1b260d-osd--block--ed04aa0b--4f5a--523b--a5d2--05072a1b260d', 'dm-uuid-LVM-viMZkgfP3jZUgTAo0bSRpd40fLtFkJJmO8XW5mmi1aRChVrq8X3DDkwGl8XRb5at'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--45b09425--beda--50b4--a83b--1fc50b5adca4-osd--block--45b09425--beda--50b4--a83b--1fc50b5adca4', 'dm-uuid-LVM-XzfTGfmLHoftdDBti3ot8l7sbRObO8yALj7bfZRysK7QQ9WjzIgwF2E0nO1Pc63W'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part1', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part14', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part15', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part16', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--ed04aa0b--4f5a--523b--a5d2--05072a1b260d-osd--block--ed04aa0b--4f5a--523b--a5d2--05072a1b260d'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-nra7ru-O0w6-dM26-lz5o-2arZ-rXsE-lANagX', 'scsi-0QEMU_QEMU_HARDDISK_4e74a544-8b3a-4a05-98a1-b7cc13f13c18', 'scsi-SQEMU_QEMU_HARDDISK_4e74a544-8b3a-4a05-98a1-b7cc13f13c18'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--ede169e1--d974--55ed--a938--c303cb89cb5b-osd--block--ede169e1--d974--55ed--a938--c303cb89cb5b', 'dm-uuid-LVM-bEBXWAqtQdfT6xKYbL0dF10We9CLdKgKh0pf0vpR2lQnL405kp4ndXs76m3o0J5Y'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--45b09425--beda--50b4--a83b--1fc50b5adca4-osd--block--45b09425--beda--50b4--a83b--1fc50b5adca4'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-XQTe4H-QTjK-DLSs-Qjj6-RKjC-da4T-xXChra', 'scsi-0QEMU_QEMU_HARDDISK_666e73fe-87f5-4f65-9ff2-95a038c8b18d', 'scsi-SQEMU_QEMU_HARDDISK_666e73fe-87f5-4f65-9ff2-95a038c8b18d'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ae4aabb-0174-4999-a1d0-4bf78cbb2965', 'scsi-SQEMU_QEMU_HARDDISK_8ae4aabb-0174-4999-a1d0-4bf78cbb2965'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part1', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part14', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part15', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part16', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a-osd--block--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a', 'dm-uuid-LVM-G1LwYmzPCvwLZRWX1y5oYSHlzZdIxlOhfbzzvktDUwfRIgHaiaw6r0MV8HVh1TLk'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-14-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--64dd100a--6bcc--5763--b557--0732c79bfd55-osd--block--64dd100a--6bcc--5763--b557--0732c79bfd55'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-HiPrG3-9qzo-WJZ0-1Smx-Bi1t-ReBz-6cdxta', 'scsi-0QEMU_QEMU_HARDDISK_f08311eb-e2ad-4eb7-8da9-ac6d899a2409', 'scsi-SQEMU_QEMU_HARDDISK_f08311eb-e2ad-4eb7-8da9-ac6d899a2409'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--5368e13a--d21f--5912--91bd--da537815d9e8-osd--block--5368e13a--d21f--5912--91bd--da537815d9e8'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-m0F2nF-j9UB-mZfT-bRv8-xW1c-EVch-ZfDvHt', 'scsi-0QEMU_QEMU_HARDDISK_e03b171b-b829-41a0-a0b6-cc46d85f013d', 'scsi-SQEMU_QEMU_HARDDISK_e03b171b-b829-41a0-a0b6-cc46d85f013d'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_9fa3c153-0e2a-45ef-8e48-e0095eb937ad', 'scsi-SQEMU_QEMU_HARDDISK_9fa3c153-0e2a-45ef-8e48-e0095eb937ad'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-21-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part1', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part14', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part15', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part16', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--ede169e1--d974--55ed--a938--c303cb89cb5b-osd--block--ede169e1--d974--55ed--a938--c303cb89cb5b'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-HVhEyD-n8kM-PvHm-CUDD-tuRZ-igjC-Nzour2', 'scsi-0QEMU_QEMU_HARDDISK_f76ae8f0-65cc-4a62-b19c-23f8a40fbde6', 'scsi-SQEMU_QEMU_HARDDISK_f76ae8f0-65cc-4a62-b19c-23f8a40fbde6'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a-osd--block--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-YifJdB-JYWE-HhRV-mF1g-0k83-hzWG-ctJZWy', 'scsi-0QEMU_QEMU_HARDDISK_d7bb81ba-6491-4b5a-bb82-f0b49314eac0', 'scsi-SQEMU_QEMU_HARDDISK_d7bb81ba-6491-4b5a-bb82-f0b49314eac0'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_cdffe5f1-7f0d-460e-a1db-d6846d226144', 'scsi-SQEMU_QEMU_HARDDISK_cdffe5f1-7f0d-460e-a1db-d6846d226144'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-28-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5', 'scsi-SQEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part1', 'scsi-SQEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part14', 'scsi-SQEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part15', 'scsi-SQEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part16', 'scsi-SQEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f', 'scsi-SQEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part1', 'scsi-SQEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part14', 'scsi-SQEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part15', 'scsi-SQEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part16', 'scsi-SQEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-23-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-10-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32', 'scsi-SQEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part1', 'scsi-SQEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part14', 'scsi-SQEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part15', 'scsi-SQEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part16', 'scsi-SQEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-19-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nSaturday 21 March 2026  00:45:16 +0000 (0:00:01.861)       0:00:37.830 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nSaturday 21 March 2026  00:45:17 +0000 (0:00:01.209)       0:00:39.040 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nSaturday 21 March 2026  00:45:18 +0000 (0:00:00.771)       0:00:39.811 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nSaturday 21 March 2026  00:45:19 +0000 (0:00:00.736)       0:00:40.547 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nSaturday 21 March 2026  00:45:19 +0000 (0:00:00.535)       0:00:41.082 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nSaturday 21 March 2026  00:45:20 +0000 (0:00:01.204)       0:00:42.287 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nSaturday 21 March 2026  00:45:21 +0000 (0:00:01.089)       0:00:43.376 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nSaturday 21 March 2026  00:45:25 +0000 (0:00:03.591)       0:00:46.968 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nSaturday 21 March 2026  00:45:27 +0000 (0:00:01.868)       0:00:48.836 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nSaturday 21 March 2026  00:45:29 +0000 (0:00:01.666)       0:00:50.503 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nSaturday 21 March 2026  00:45:29 +0000 (0:00:00.312)       0:00:50.815 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nSaturday 21 March 2026  00:45:29 +0000 (0:00:00.440)       0:00:51.256 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nSaturday 21 March 2026  00:45:30 +0000 (0:00:00.376)       0:00:51.632 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nSaturday 21 March 2026  00:45:31 +0000 (0:00:01.114)       0:00:52.747 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nSaturday 21 March 2026  00:45:31 +0000 (0:00:00.435)       0:00:53.183 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nSaturday 21 March 2026  00:45:32 +0000 (0:00:00.491)       0:00:53.675 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nSaturday 21 March 2026  00:45:32 +0000 (0:00:00.405)       0:00:54.081 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nSaturday 21 March 2026  00:45:32 +0000 (0:00:00.318)       0:00:54.400 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nSaturday 21 March 2026  00:45:33 +0000 (0:00:00.712)       0:00:55.113 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nSaturday 21 March 2026  00:45:34 +0000 (0:00:01.133)       0:00:56.246 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nSaturday 21 March 2026  00:45:36 +0000 (0:00:02.014)       0:00:58.260 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nSaturday 21 March 2026  00:45:37 +0000 (0:00:01.144)       0:00:59.404 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nSaturday 21 March 2026  00:45:39 +0000 (0:00:01.251)       0:01:00.656 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nSaturday 21 March 2026  00:45:40 +0000 (0:00:01.522)       0:01:02.179 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nSaturday 21 March 2026  00:45:41 +0000 (0:00:00.984)       0:01:03.163 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nSaturday 21 March 2026  00:45:42 +0000 (0:00:01.222)       0:01:04.386 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nSaturday 21 March 2026  00:45:44 +0000 (0:00:01.779)       0:01:06.165 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nSaturday 21 March 2026  00:45:45 +0000 (0:00:00.955)       0:01:07.121 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nSaturday 21 March 2026  00:45:46 +0000 (0:00:00.983)       0:01:08.105 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nSaturday 21 March 2026  00:45:47 +0000 (0:00:00.813)       0:01:08.919 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nSaturday 21 March 2026  00:45:49 +0000 (0:00:02.087)       0:01:11.007 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nSaturday 21 March 2026  00:45:51 +0000 (0:00:02.101)       0:01:13.108 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nSaturday 21 March 2026  00:45:52 +0000 (0:00:01.278)       0:01:14.386 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nSaturday 21 March 2026  00:45:53 +0000 (0:00:00.918)       0:01:15.304 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nSaturday 21 March 2026  00:45:54 +0000 (0:00:00.874)       0:01:16.179 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nSaturday 21 March 2026  00:45:55 +0000 (0:00:00.997)       0:01:17.177 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nSaturday 21 March 2026  00:45:56 +0000 (0:00:01.029)       0:01:18.207 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nSaturday 21 March 2026  00:45:57 +0000 (0:00:00.687)       0:01:18.895 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nSaturday 21 March 2026  00:45:58 +0000 (0:00:01.004)       0:01:19.899 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nSaturday 21 March 2026  00:45:59 +0000 (0:00:00.868)       0:01:20.768 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nSaturday 21 March 2026  00:46:00 +0000 (0:00:00.906)       0:01:21.674 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Generate systemd ceph target file] ***************\nSaturday 21 March 2026  00:46:01 +0000 (0:00:01.318)       0:01:22.993 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Enable ceph.target] ******************************\nSaturday 21 March 2026  00:46:02 +0000 (0:00:01.346)       0:01:24.339 ******** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include prerequisites.yml] ***********************\nSaturday 21 March 2026  00:46:05 +0000 (0:00:02.188)       0:01:26.528 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-container-common : Stop lvmetad] ************************************\nSaturday 21 March 2026  00:46:06 +0000 (0:00:01.073)       0:01:27.602 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Disable and mask lvmetad service] ****************\nSaturday 21 March 2026  00:46:06 +0000 (0:00:00.542)       0:01:28.145 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Remove ceph udev rules] **************************\nSaturday 21 March 2026  00:46:07 +0000 (0:00:00.663)       0:01:28.808 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\nTASK [ceph-container-common : Ensure tmpfiles.d is present] ********************\nSaturday 21 March 2026  00:46:08 +0000 (0:00:01.225)       0:01:30.034 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Restore certificates selinux context] ************\nSaturday 21 March 2026  00:46:09 +0000 (0:00:00.974)       0:01:31.009 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Install python3 on osd nodes] ********************\nSaturday 21 March 2026  00:46:10 +0000 (0:00:00.513)       0:01:31.523 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include registry.yml] ****************************\nSaturday 21 March 2026  00:46:10 +0000 (0:00:00.668)       0:01:32.191 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include fetch_image.yml] *************************\nSaturday 21 March 2026  00:46:11 +0000 (0:00:00.492)       0:01:32.684 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-container-common : Pulling Ceph container image] ********************\nSaturday 21 March 2026  00:46:12 +0000 (0:00:01.094)       0:01:33.779 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***\nSaturday 21 March 2026  00:47:10 +0000 (0:00:58.605)       0:02:32.384 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Pulling node-exporter container image] ***********\nSaturday 21 March 2026  00:47:11 +0000 (0:00:00.579)       0:02:32.963 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Export local ceph dev image] *********************\nSaturday 21 March 2026  00:47:12 +0000 (0:00:00.640)       0:02:33.604 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Copy ceph dev image file] ************************\nSaturday 21 March 2026  00:47:12 +0000 (0:00:00.150)       0:02:33.754 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Load ceph dev image] *****************************\nSaturday 21 March 2026  00:47:12 +0000 (0:00:00.522)       0:02:34.277 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Remove tmp ceph dev image file] ******************\nSaturday 21 March 2026  00:47:13 +0000 (0:00:00.742)       0:02:35.020 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Get ceph version] ********************************\nSaturday 21 March 2026  00:47:14 +0000 (0:00:00.515)       0:02:35.535 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***\nSaturday 21 March 2026  00:47:16 +0000 (0:00:02.715)       0:02:38.251 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include release.yml] *****************************\nSaturday 21 March 2026  00:47:17 +0000 (0:00:00.662)       0:02:38.913 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release jewel] *********************\nSaturday 21 March 2026  00:47:18 +0000 (0:00:01.095)       0:02:40.009 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release kraken] ********************\nSaturday 21 March 2026  00:47:19 +0000 (0:00:00.533)       0:02:40.542 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release luminous] ******************\nSaturday 21 March 2026  00:47:19 +0000 (0:00:00.784)       0:02:41.326 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release mimic] *********************\nSaturday 21 March 2026  00:47:20 +0000 (0:00:00.559)       0:02:41.886 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release nautilus] ******************\nSaturday 21 March 2026  00:47:21 +0000 (0:00:00.741)       0:02:42.628 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release octopus] *******************\nSaturday 21 March 2026  00:47:21 +0000 (0:00:00.517)       0:02:43.146 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release pacific] *******************\nSaturday 21 March 2026  00:47:22 +0000 (0:00:00.685)       0:02:43.831 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release quincy] ********************\nSaturday 21 March 2026  00:47:22 +0000 (0:00:00.534)       0:02:44.366 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release reef] **********************\nSaturday 21 March 2026  00:47:23 +0000 (0:00:00.683)       0:02:45.050 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************\nSaturday 21 March 2026  00:47:25 +0000 (0:00:01.434)       0:02:46.485 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-config : Create ceph initial directories] ***************************\nSaturday 21 March 2026  00:47:26 +0000 (0:00:01.148)       0:02:47.633 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/log/ceph)\u001b[0m\n\nTASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************\nSaturday 21 March 2026  00:47:32 +0000 (0:00:06.694)       0:02:54.328 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-config : Create rados gateway instance directories] *****************\nSaturday 21 March 2026  00:47:34 +0000 (0:00:01.234)       0:02:55.562 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Generate environment file] *********************************\nSaturday 21 March 2026  00:47:34 +0000 (0:00:00.628)       0:02:56.191 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Reset num_osds] ********************************************\nSaturday 21 March 2026  00:47:35 +0000 (0:00:01.177)       0:02:57.369 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Count number of osds for lvm scenario] *********************\nSaturday 21 March 2026  00:47:36 +0000 (0:00:00.562)       0:02:57.931 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Look up for ceph-volume rejected devices] ******************\nSaturday 21 March 2026  00:47:37 +0000 (0:00:00.843)       0:02:58.775 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact rejected_devices] *********************************\nSaturday 21 March 2026  00:47:38 +0000 (0:00:00.666)       0:02:59.442 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact _devices] *****************************************\nSaturday 21 March 2026  00:47:38 +0000 (0:00:00.846)       0:03:00.288 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***\nSaturday 21 March 2026  00:47:40 +0000 (0:00:01.439)       0:03:01.728 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***\nSaturday 21 March 2026  00:47:40 +0000 (0:00:00.644)       0:03:02.372 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***\nSaturday 21 March 2026  00:47:41 +0000 (0:00:00.706)       0:03:03.078 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***\nSaturday 21 March 2026  00:47:42 +0000 (0:00:00.789)       0:03:03.868 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds (add existing osds)] *********************\nSaturday 21 March 2026  00:47:45 +0000 (0:00:03.449)       0:03:07.317 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact _osd_memory_target] *******************************\nSaturday 21 March 2026  00:47:46 +0000 (0:00:00.543)       0:03:07.861 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set osd_memory_target to cluster host config] **************\nSaturday 21 March 2026  00:47:47 +0000 (0:00:00.799)       0:03:08.661 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Render rgw configs] ****************************************\nSaturday 21 March 2026  00:47:47 +0000 (0:00:00.582)       0:03:09.243 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set config to cluster] *************************************\nSaturday 21 March 2026  00:47:48 +0000 (0:00:00.708)       0:03:09.952 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.13:8081'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.14:8081'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.15:8081'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set rgw configs to file] ***********************************\nSaturday 21 March 2026  00:47:49 +0000 (0:00:00.580)       0:03:10.532 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Create ceph conf directory] ********************************\nSaturday 21 March 2026  00:47:50 +0000 (0:00:00.919)       0:03:11.452 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nSaturday 21 March 2026  00:47:50 +0000 (0:00:00.728)       0:03:12.181 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nSaturday 21 March 2026  00:47:51 +0000 (0:00:01.079)       0:03:13.260 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nSaturday 21 March 2026  00:47:52 +0000 (0:00:00.741)       0:03:14.002 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nSaturday 21 March 2026  00:47:53 +0000 (0:00:00.982)       0:03:14.985 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nSaturday 21 March 2026  00:47:54 +0000 (0:00:00.695)       0:03:15.680 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nSaturday 21 March 2026  00:47:54 +0000 (0:00:00.540)       0:03:16.221 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nSaturday 21 March 2026  00:47:55 +0000 (0:00:00.582)       0:03:16.803 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nSaturday 21 March 2026  00:47:56 +0000 (0:00:00.700)       0:03:17.503 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nSaturday 21 March 2026  00:47:56 +0000 (0:00:00.490)       0:03:17.993 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=0) \u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Generate Ceph file] ****************************************\nSaturday 21 March 2026  00:47:57 +0000 (0:00:01.344)       0:03:19.338 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nSaturday 21 March 2026  00:48:00 +0000 (0:00:02.535)       0:03:21.873 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mons handler] **********************************\nSaturday 21 March 2026  00:48:01 +0000 (0:00:01.381)       0:03:23.254 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mons.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mon_handler_called before restart] ********\nSaturday 21 March 2026  00:48:02 +0000 (0:00:00.778)       0:03:24.033 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mon restart script] ***********************\nSaturday 21 March 2026  00:48:02 +0000 (0:00:00.331)       0:03:24.364 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mon daemon(s)] ********************\nSaturday 21 March 2026  00:48:04 +0000 (0:00:01.247)       0:03:25.612 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mon_handler_called after restart] *********\nSaturday 21 March 2026  00:48:04 +0000 (0:00:00.787)       0:03:26.399 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Osds handler] **********************************\nSaturday 21 March 2026  00:48:05 +0000 (0:00:00.281)       0:03:26.680 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set_fact trigger_restart] **********************\nSaturday 21 March 2026  00:48:06 +0000 (0:00:01.206)       0:03:27.887 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _osd_handler_called before restart] ********\nSaturday 21 March 2026  00:48:06 +0000 (0:00:00.405)       0:03:28.292 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Unset noup flag] *******************************\nSaturday 21 March 2026  00:48:07 +0000 (0:00:00.674)       0:03:28.967 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy osd restart script] ***********************\nSaturday 21 March 2026  00:48:07 +0000 (0:00:00.218)       0:03:29.186 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Get pool list] *********************************\nSaturday 21 March 2026  00:48:08 +0000 (0:00:00.286)       0:03:29.473 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Get balancer module status] ********************\nSaturday 21 March 2026  00:48:08 +0000 (0:00:00.258)       0:03:29.731 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set_fact pools_pgautoscaler_mode] **************\nSaturday 21 March 2026  00:48:08 +0000 (0:00:00.174)       0:03:29.906 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Disable balancer] ******************************\nSaturday 21 March 2026  00:48:08 +0000 (0:00:00.105)       0:03:30.012 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Disable pg autoscale on pools] *****************\nSaturday 21 March 2026  00:48:08 +0000 (0:00:00.220)       0:03:30.233 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph osds daemon(s)] *******************\nSaturday 21 March 2026  00:48:09 +0000 (0:00:00.216)       0:03:30.449 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _osd_handler_called after restart] *********\nSaturday 21 March 2026  00:48:09 +0000 (0:00:00.550)       0:03:31.000 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Re-enable pg autoscale on pools] ***************\nSaturday 21 March 2026  00:48:10 +0000 (0:00:00.441)       0:03:31.441 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Re-enable balancer] ****************************\nSaturday 21 March 2026  00:48:10 +0000 (0:00:00.225)       0:03:31.667 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mdss handler] **********************************\nSaturday 21 March 2026  00:48:10 +0000 (0:00:00.232)       0:03:31.900 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mdss.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mds_handler_called before restart] ********\nSaturday 21 March 2026  00:48:11 +0000 (0:00:00.813)       0:03:32.713 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mds restart script] ***********************\nSaturday 21 March 2026  00:48:11 +0000 (0:00:00.440)       0:03:33.154 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mds daemon(s)] ********************\nSaturday 21 March 2026  00:48:12 +0000 (0:00:00.994)       0:03:34.148 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mds_handler_called after restart] *********\nSaturday 21 March 2026  00:48:13 +0000 (0:00:00.536)       0:03:34.685 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rgws handler] **********************************\nSaturday 21 March 2026  00:48:13 +0000 (0:00:00.258)       0:03:34.943 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_rgws.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _rgw_handler_called before restart] ********\nSaturday 21 March 2026  00:48:14 +0000 (0:00:00.905)       0:03:35.849 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy rgw restart script] ***********************\nSaturday 21 March 2026  00:48:14 +0000 (0:00:00.316)       0:03:36.165 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph rgw daemon(s)] ********************\nSaturday 21 March 2026  00:48:15 +0000 (0:00:01.221)       0:03:37.387 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _rgw_handler_called after restart] *********\nSaturday 21 March 2026  00:48:16 +0000 (0:00:00.554)       0:03:37.941 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rbdmirrors handler] ****************************\nSaturday 21 March 2026  00:48:16 +0000 (0:00:00.349)       0:03:38.291 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mgrs handler] **********************************\nSaturday 21 March 2026  00:48:17 +0000 (0:00:00.605)       0:03:38.896 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mgrs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mgr_handler_called before restart] ********\nSaturday 21 March 2026  00:48:18 +0000 (0:00:00.977)       0:03:39.873 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mgr restart script] ***********************\nSaturday 21 March 2026  00:48:18 +0000 (0:00:00.277)       0:03:40.151 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mgr daemon(s)] ********************\nSaturday 21 March 2026  00:48:20 +0000 (0:00:01.376)       0:03:41.527 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mgr_handler_called after restart] *********\nSaturday 21 March 2026  00:48:20 +0000 (0:00:00.528)       0:03:42.056 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role ceph-mon] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nSaturday 21 March 2026  00:48:21 +0000 (0:00:00.490)       0:03:42.546 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nSaturday 21 March 2026  00:48:21 +0000 (0:00:00.601)       0:03:43.148 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nSaturday 21 March 2026  00:48:22 +0000 (0:00:00.494)       0:03:43.642 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nSaturday 21 March 2026  00:48:22 +0000 (0:00:00.621)       0:03:44.263 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nSaturday 21 March 2026  00:48:23 +0000 (0:00:00.285)       0:03:44.549 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nSaturday 21 March 2026  00:48:23 +0000 (0:00:00.452)       0:03:45.002 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nSaturday 21 March 2026  00:48:23 +0000 (0:00:00.307)       0:03:45.310 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nSaturday 21 March 2026  00:48:24 +0000 (0:00:00.615)       0:03:45.925 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nSaturday 21 March 2026  00:48:24 +0000 (0:00:00.284)       0:03:46.210 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nSaturday 21 March 2026  00:48:25 +0000 (0:00:00.495)       0:03:46.705 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nSaturday 21 March 2026  00:48:26 +0000 (0:00:00.795)       0:03:47.501 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nSaturday 21 March 2026  00:48:26 +0000 (0:00:00.712)       0:03:48.213 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nSaturday 21 March 2026  00:48:27 +0000 (0:00:00.266)       0:03:48.480 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nSaturday 21 March 2026  00:48:27 +0000 (0:00:00.441)       0:03:48.922 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nSaturday 21 March 2026  00:48:27 +0000 (0:00:00.320)       0:03:49.243 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nSaturday 21 March 2026  00:48:28 +0000 (0:00:00.253)       0:03:49.496 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nSaturday 21 March 2026  00:48:28 +0000 (0:00:00.257)       0:03:49.754 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nSaturday 21 March 2026  00:48:28 +0000 (0:00:00.448)       0:03:50.203 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nSaturday 21 March 2026  00:48:29 +0000 (0:00:00.261)       0:03:50.464 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nSaturday 21 March 2026  00:48:29 +0000 (0:00:00.292)       0:03:50.757 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nSaturday 21 March 2026  00:48:29 +0000 (0:00:00.296)       0:03:51.054 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set_fact container_exec_cmd] **********************************\nSaturday 21 March 2026  00:48:30 +0000 (0:00:00.749)       0:03:51.803 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include deploy_monitors.yml] **********************************\nSaturday 21 March 2026  00:48:30 +0000 (0:00:00.287)       0:03:52.091 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/deploy_monitors.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mon : Check if monitor initial keyring already exists] **************\nSaturday 21 March 2026  00:48:31 +0000 (0:00:00.478)       0:03:52.569 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Generate monitor initial keyring] *****************************\nSaturday 21 March 2026  00:48:31 +0000 (0:00:00.375)       0:03:52.944 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [ceph-mon : Set_fact _initial_mon_key_success] ****************************\nSaturday 21 March 2026  00:48:32 +0000 (0:00:01.037)       0:03:53.982 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Get initial keyring when it already exists] *******************\nSaturday 21 March 2026  00:48:33 +0000 (0:00:00.439)       0:03:54.421 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Create monitor initial keyring] *******************************\nSaturday 21 March 2026  00:48:33 +0000 (0:00:00.358)       0:03:54.780 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Copy the initial key in /etc/ceph (for containers)] ***********\nSaturday 21 March 2026  00:48:34 +0000 (0:00:01.014)       0:03:55.794 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Create monitor directory] *************************************\nSaturday 21 March 2026  00:48:35 +0000 (0:00:01.002)       0:03:56.797 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Recursively fix ownership of monitor directory] ***************\nSaturday 21 March 2026  00:48:36 +0000 (0:00:00.713)       0:03:57.510 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Create admin keyring] *****************************************\nSaturday 21 March 2026  00:48:36 +0000 (0:00:00.680)       0:03:58.191 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Slurp admin keyring] ******************************************\nSaturday 21 March 2026  00:48:37 +0000 (0:00:01.156)       0:03:59.347 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Copy admin keyring over to mons] ******************************\nSaturday 21 March 2026  00:48:38 +0000 (0:00:00.738)       0:04:00.086 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0 -> {{ item }}]\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> {{ item }}]\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> {{ item }}]\u001b[0m\n\nTASK [ceph-mon : Import admin keyring into mon keyring] ************************\nSaturday 21 March 2026  00:48:42 +0000 (0:00:04.013)       0:04:04.099 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set_fact ceph-mon container command] **************************\nSaturday 21 March 2026  00:48:44 +0000 (0:00:01.350)       0:04:05.450 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set_fact monmaptool container command] ************************\nSaturday 21 March 2026  00:48:44 +0000 (0:00:00.417)       0:04:05.867 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Generate initial monmap] **************************************\nSaturday 21 March 2026  00:48:44 +0000 (0:00:00.439)       0:04:06.307 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Ceph monitor mkfs with keyring] *******************************\nSaturday 21 March 2026  00:48:47 +0000 (0:00:02.405)       0:04:08.712 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Ceph monitor mkfs without keyring] ****************************\nSaturday 21 March 2026  00:48:49 +0000 (0:00:01.738)       0:04:10.451 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include start_monitor.yml] ************************************\nSaturday 21 March 2026  00:48:49 +0000 (0:00:00.279)       0:04:10.730 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/start_monitor.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mon : Ensure systemd service override directory exists] *************\nSaturday 21 March 2026  00:48:49 +0000 (0:00:00.421)       0:04:11.152 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Add ceph-mon systemd service overrides] ***********************\nSaturday 21 March 2026  00:48:50 +0000 (0:00:00.968)       0:04:12.121 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Include_tasks systemd.yml] ************************************\nSaturday 21 March 2026  00:48:51 +0000 (0:00:00.415)       0:04:12.536 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/systemd.yml for testbed-node-0, testbed-node-2, testbed-node-1\u001b[0m\n\nTASK [ceph-mon : Generate systemd unit file for mon container] *****************\nSaturday 21 March 2026  00:48:52 +0000 (0:00:00.922)       0:04:13.458 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Generate systemd ceph-mon target file] ************************\nSaturday 21 March 2026  00:48:53 +0000 (0:00:01.925)       0:04:15.383 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Enable ceph-mon.target] ***************************************\nSaturday 21 March 2026  00:48:55 +0000 (0:00:01.481)       0:04:16.865 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Start the monitor service] ************************************\nSaturday 21 March 2026  00:48:57 +0000 (0:00:01.710)       0:04:18.575 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include_tasks ceph_keys.yml] **********************************\nSaturday 21 March 2026  00:48:59 +0000 (0:00:01.942)       0:04:20.517 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/ceph_keys.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mon : Waiting for the monitor(s) to form the quorum...] *************\nSaturday 21 March 2026  00:48:59 +0000 (0:00:00.651)       0:04:21.169 ******** \n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for the monitor(s) to form the quorum... (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Fetch ceph initial keys] **************************************\nSaturday 21 March 2026  00:49:21 +0000 (0:00:21.519)       0:04:42.689 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Include secure_cluster.yml] ***********************************\nSaturday 21 March 2026  00:49:30 +0000 (0:00:08.912)       0:04:51.601 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set cluster configs] ******************************************\nSaturday 21 March 2026  00:49:30 +0000 (0:00:00.255)       0:04:51.857 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__8519e6c23c9c83674cbc6ea19a79c2df830b7b57'}}, {'key': 'public_network', 'value': '192.168.16.0/20'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__8519e6c23c9c83674cbc6ea19a79c2df830b7b57'}}, {'key': 'cluster_network', 'value': '192.168.16.0/20'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__8519e6c23c9c83674cbc6ea19a79c2df830b7b57'}}, {'key': 'osd_pool_default_crush_rule', 'value': -1}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__8519e6c23c9c83674cbc6ea19a79c2df830b7b57'}}, {'key': 'ms_bind_ipv6', 'value': 'False'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__8519e6c23c9c83674cbc6ea19a79c2df830b7b57'}}, {'key': 'ms_bind_ipv4', 'value': 'True'}])\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__8519e6c23c9c83674cbc6ea19a79c2df830b7b57'}}, {'key': 'osd_crush_chooseleaf_type', 'value': '__omit_place_holder__8519e6c23c9c83674cbc6ea19a79c2df830b7b57'}]) \u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nSaturday 21 March 2026  00:49:45 +0000 (0:00:14.955)       0:05:06.812 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mons handler] **********************************\nSaturday 21 March 2026  00:49:45 +0000 (0:00:00.322)       0:05:07.135 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mons.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mon_handler_called before restart] ********\nSaturday 21 March 2026  00:49:46 +0000 (0:00:00.477)       0:05:07.613 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mon restart script] ***********************\nSaturday 21 March 2026  00:49:46 +0000 (0:00:00.424)       0:05:08.037 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mon daemon(s)] ********************\nSaturday 21 March 2026  00:49:46 +0000 (0:00:00.277)       0:05:08.314 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mon_handler_called after restart] *********\nSaturday 21 March 2026  00:49:47 +0000 (0:00:00.537)       0:05:08.851 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role ceph-mgr] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nSaturday 21 March 2026  00:49:48 +0000 (0:00:00.632)       0:05:09.484 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nSaturday 21 March 2026  00:49:48 +0000 (0:00:00.431)       0:05:09.915 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nSaturday 21 March 2026  00:49:48 +0000 (0:00:00.464)       0:05:10.379 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nSaturday 21 March 2026  00:49:49 +0000 (0:00:00.866)       0:05:11.246 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nSaturday 21 March 2026  00:49:50 +0000 (0:00:00.267)       0:05:11.513 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nSaturday 21 March 2026  00:49:50 +0000 (0:00:00.304)       0:05:11.817 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nSaturday 21 March 2026  00:49:50 +0000 (0:00:00.298)       0:05:12.116 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nSaturday 21 March 2026  00:49:51 +0000 (0:00:00.971)       0:05:13.087 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nSaturday 21 March 2026  00:49:51 +0000 (0:00:00.279)       0:05:13.367 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nSaturday 21 March 2026  00:49:52 +0000 (0:00:00.262)       0:05:13.630 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nSaturday 21 March 2026  00:49:52 +0000 (0:00:00.691)       0:05:14.321 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nSaturday 21 March 2026  00:49:53 +0000 (0:00:00.860)       0:05:15.181 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nSaturday 21 March 2026  00:49:54 +0000 (0:00:00.274)       0:05:15.456 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nSaturday 21 March 2026  00:49:54 +0000 (0:00:00.275)       0:05:15.731 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nSaturday 21 March 2026  00:49:54 +0000 (0:00:00.246)       0:05:15.978 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nSaturday 21 March 2026  00:49:54 +0000 (0:00:00.403)       0:05:16.381 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nSaturday 21 March 2026  00:49:55 +0000 (0:00:00.256)       0:05:16.638 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nSaturday 21 March 2026  00:49:55 +0000 (0:00:00.250)       0:05:16.888 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nSaturday 21 March 2026  00:49:55 +0000 (0:00:00.272)       0:05:17.161 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nSaturday 21 March 2026  00:49:55 +0000 (0:00:00.258)       0:05:17.420 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nSaturday 21 March 2026  00:49:56 +0000 (0:00:00.490)       0:05:17.910 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Set_fact container_exec_cmd] **********************************\nSaturday 21 March 2026  00:49:56 +0000 (0:00:00.445)       0:05:18.356 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-mgr : Include common.yml] *******************************************\nSaturday 21 March 2026  00:49:57 +0000 (0:00:00.700)       0:05:19.056 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/common.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mgr : Create mgr directory] *****************************************\nSaturday 21 March 2026  00:49:58 +0000 (0:00:00.609)       0:05:19.666 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Fetch ceph mgr keyring] ***************************************\nSaturday 21 March 2026  00:49:58 +0000 (0:00:00.658)       0:05:20.325 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Create ceph mgr keyring(s) on a mon node] *********************\nSaturday 21 March 2026  00:49:59 +0000 (0:00:00.264)       0:05:20.589 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0 -> {{ groups[mon_group_name][0] }}]\u001b[0m\n\nTASK [ceph-mgr : Set_fact _mgr_keys] *******************************************\nSaturday 21 March 2026  00:50:09 +0000 (0:00:10.782)       0:05:31.372 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Get keys from monitors] ***************************************\nSaturday 21 March 2026  00:50:10 +0000 (0:00:00.686)       0:05:32.058 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\nTASK [ceph-mgr : Copy ceph key(s) if needed] ***********************************\nSaturday 21 March 2026  00:50:13 +0000 (0:00:02.652)       0:05:34.711 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=None)\u001b[0m\n\nTASK [ceph-mgr : Set mgr key permissions] **************************************\nSaturday 21 March 2026  00:50:14 +0000 (0:00:01.345)       0:05:36.056 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Append dashboard modules to ceph_mgr_modules] *****************\nSaturday 21 March 2026  00:50:15 +0000 (0:00:00.788)       0:05:36.844 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include pre_requisite.yml] ************************************\nSaturday 21 March 2026  00:50:15 +0000 (0:00:00.547)       0:05:37.392 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include start_mgr.yml] ****************************************\nSaturday 21 March 2026  00:50:16 +0000 (0:00:00.278)       0:05:37.670 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/start_mgr.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mgr : Ensure systemd service override directory exists] *************\nSaturday 21 March 2026  00:50:16 +0000 (0:00:00.454)       0:05:38.124 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Add ceph-mgr systemd service overrides] ***********************\nSaturday 21 March 2026  00:50:16 +0000 (0:00:00.275)       0:05:38.399 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include_tasks systemd.yml] ************************************\nSaturday 21 March 2026  00:50:17 +0000 (0:00:00.473)       0:05:38.873 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/systemd.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mgr : Generate systemd unit file] ***********************************\nSaturday 21 March 2026  00:50:17 +0000 (0:00:00.433)       0:05:39.306 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Generate systemd ceph-mgr target file] ************************\nSaturday 21 March 2026  00:50:19 +0000 (0:00:01.191)       0:05:40.498 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Enable ceph-mgr.target] ***************************************\nSaturday 21 March 2026  00:50:20 +0000 (0:00:01.392)       0:05:41.890 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Systemd start mgr] ********************************************\nSaturday 21 March 2026  00:50:22 +0000 (0:00:01.970)       0:05:43.860 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include mgr_modules.yml] **************************************\nSaturday 21 March 2026  00:50:24 +0000 (0:00:01.986)       0:05:45.847 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/mgr_modules.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-mgr : Wait for all mgr to be up] ************************************\nSaturday 21 March 2026  00:50:24 +0000 (0:00:00.389)       0:05:46.237 ******** \n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (30 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (29 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (28 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (27 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (26 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (25 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-mgr : Get enabled modules from ceph-mgr] ****************************\nSaturday 21 March 2026  00:51:01 +0000 (0:00:36.711)       0:06:22.949 ******** \n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-mgr : Set _ceph_mgr_modules fact (convert _ceph_mgr_modules.stdout to a dict)] ***\nSaturday 21 March 2026  00:51:02 +0000 (0:00:01.236)       0:06:24.185 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Set _disabled_ceph_mgr_modules fact] **************************\nSaturday 21 March 2026  00:51:03 +0000 (0:00:00.300)       0:06:24.486 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Disable ceph mgr enabled modules] *****************************\nSaturday 21 March 2026  00:51:03 +0000 (0:00:00.115)       0:06:24.601 ******** \n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=iostat)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=nfs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=restful)\u001b[0m\n\nTASK [ceph-mgr : Add modules to ceph-mgr] **************************************\nSaturday 21 March 2026  00:51:09 +0000 (0:00:06.385)       0:06:30.986 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item=balancer) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=dashboard)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=prometheus)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=status) \u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nSaturday 21 March 2026  00:51:14 +0000 (0:00:04.811)       0:06:35.798 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mgrs handler] **********************************\nSaturday 21 March 2026  00:51:15 +0000 (0:00:01.012)       0:06:36.811 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mgrs.yml for testbed-node-1, testbed-node-0, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mgr_handler_called before restart] ********\nSaturday 21 March 2026  00:51:16 +0000 (0:00:00.614)       0:06:37.425 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mgr restart script] ***********************\nSaturday 21 March 2026  00:51:16 +0000 (0:00:00.320)       0:06:37.745 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mgr daemon(s)] ********************\nSaturday 21 March 2026  00:51:17 +0000 (0:00:01.368)       0:06:39.114 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mgr_handler_called after restart] *********\nSaturday 21 March 2026  00:51:18 +0000 (0:00:00.575)       0:06:39.689 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role ceph-osd] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nSaturday 21 March 2026  00:51:18 +0000 (0:00:00.491)       0:06:40.181 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nSaturday 21 March 2026  00:51:19 +0000 (0:00:00.574)       0:06:40.755 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-5, testbed-node-4\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nSaturday 21 March 2026  00:51:19 +0000 (0:00:00.489)       0:06:41.245 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nSaturday 21 March 2026  00:51:20 +0000 (0:00:00.294)       0:06:41.540 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nSaturday 21 March 2026  00:51:20 +0000 (0:00:00.800)       0:06:42.341 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nSaturday 21 March 2026  00:51:21 +0000 (0:00:00.631)       0:06:42.973 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nSaturday 21 March 2026  00:51:22 +0000 (0:00:00.727)       0:06:43.701 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nSaturday 21 March 2026  00:51:22 +0000 (0:00:00.252)       0:06:43.953 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nSaturday 21 March 2026  00:51:22 +0000 (0:00:00.406)       0:06:44.359 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nSaturday 21 March 2026  00:51:23 +0000 (0:00:00.268)       0:06:44.628 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nSaturday 21 March 2026  00:51:23 +0000 (0:00:00.626)       0:06:45.254 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nSaturday 21 March 2026  00:51:24 +0000 (0:00:00.588)       0:06:45.843 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nSaturday 21 March 2026  00:51:24 +0000 (0:00:00.469)       0:06:46.312 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nSaturday 21 March 2026  00:51:25 +0000 (0:00:00.249)       0:06:46.562 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nSaturday 21 March 2026  00:51:25 +0000 (0:00:00.285)       0:06:46.848 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nSaturday 21 March 2026  00:51:25 +0000 (0:00:00.280)       0:06:47.128 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nSaturday 21 March 2026  00:51:26 +0000 (0:00:00.487)       0:06:47.616 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nSaturday 21 March 2026  00:51:26 +0000 (0:00:00.283)       0:06:47.899 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nSaturday 21 March 2026  00:51:26 +0000 (0:00:00.246)       0:06:48.146 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nSaturday 21 March 2026  00:51:26 +0000 (0:00:00.255)       0:06:48.402 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nSaturday 21 March 2026  00:51:27 +0000 (0:00:00.525)       0:06:48.928 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Set_fact add_osd] *********************************************\nSaturday 21 March 2026  00:51:27 +0000 (0:00:00.458)       0:06:49.386 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Set_fact container_exec_cmd] **********************************\nSaturday 21 March 2026  00:51:28 +0000 (0:00:00.278)       0:06:49.665 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-osd : Include_tasks system_tuning.yml] ******************************\nSaturday 21 March 2026  00:51:28 +0000 (0:00:00.714)       0:06:50.380 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/system_tuning.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Create tmpfiles.d directory] **********************************\nSaturday 21 March 2026  00:51:29 +0000 (0:00:00.644)       0:06:51.024 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Disable transparent hugepage] *********************************\nSaturday 21 March 2026  00:51:29 +0000 (0:00:00.241)       0:06:51.266 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Get default vm.min_free_kbytes] *******************************\nSaturday 21 March 2026  00:51:30 +0000 (0:00:00.272)       0:06:51.538 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Set_fact vm_min_free_kbytes] **********************************\nSaturday 21 March 2026  00:51:30 +0000 (0:00:00.858)       0:06:52.397 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Apply operating system tuning] ********************************\nSaturday 21 March 2026  00:51:31 +0000 (0:00:00.351)       0:06:52.748 ******** \n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m\n\nTASK [ceph-osd : Install dependencies] *****************************************\nSaturday 21 March 2026  00:51:35 +0000 (0:00:03.956)       0:06:56.704 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Include_tasks common.yml] *************************************\nSaturday 21 March 2026  00:51:35 +0000 (0:00:00.253)       0:06:56.958 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Create bootstrap-osd and osd directories] *********************\nSaturday 21 March 2026  00:51:36 +0000 (0:00:00.612)       0:06:57.570 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/osd/)\u001b[0m\n\nTASK [ceph-osd : Get keys from monitors] ***************************************\nSaturday 21 March 2026  00:51:37 +0000 (0:00:00.955)       0:06:58.526 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-osd : Copy ceph key(s) if needed] ***********************************\nSaturday 21 March 2026  00:51:39 +0000 (0:00:02.197)       0:07:00.724 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Set noup flag] ************************************************\nSaturday 21 March 2026  00:51:40 +0000 (0:00:01.456)       0:07:02.180 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-osd : Include_tasks scenarios/lvm.yml] ******************************\nSaturday 21 March 2026  00:51:42 +0000 (0:00:02.154)       0:07:04.334 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/scenarios/lvm.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Use ceph-volume to create osds] *******************************\nSaturday 21 March 2026  00:51:43 +0000 (0:00:00.457)       0:07:04.792 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'})\u001b[0m\n\nTASK [ceph-osd : Include_tasks scenarios/lvm-batch.yml] ************************\nSaturday 21 March 2026  00:52:27 +0000 (0:00:44.172)       0:07:48.964 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Include_tasks start_osds.yml] *********************************\nSaturday 21 March 2026  00:52:28 +0000 (0:00:00.470)       0:07:49.434 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/start_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Get osd ids] **************************************************\nSaturday 21 March 2026  00:52:28 +0000 (0:00:00.438)       0:07:49.873 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Collect osd ids] **********************************************\nSaturday 21 March 2026  00:52:29 +0000 (0:00:00.662)       0:07:50.535 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Include_tasks systemd.yml] ************************************\nSaturday 21 March 2026  00:52:32 +0000 (0:00:02.920)       0:07:53.455 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Generate systemd unit file] ***********************************\nSaturday 21 March 2026  00:52:32 +0000 (0:00:00.470)       0:07:53.926 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Generate systemd ceph-osd target file] ************************\nSaturday 21 March 2026  00:52:33 +0000 (0:00:01.227)       0:07:55.153 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Enable ceph-osd.target] ***************************************\nSaturday 21 March 2026  00:52:35 +0000 (0:00:01.382)       0:07:56.535 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [ceph-osd : Ensure systemd service override directory exists] *************\nSaturday 21 March 2026  00:52:36 +0000 (0:00:01.865)       0:07:58.401 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Add ceph-osd systemd service overrides] ***********************\nSaturday 21 March 2026  00:52:37 +0000 (0:00:00.322)       0:07:58.723 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Ensure /var/lib/ceph/osd/<cluster>-<item> is present] *********\nSaturday 21 March 2026  00:52:37 +0000 (0:00:00.285)       0:07:59.009 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=3)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=5)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=2)\u001b[0m\n\nTASK [ceph-osd : Write run file in /var/lib/ceph/osd/xxxx/run] *****************\nSaturday 21 March 2026  00:52:38 +0000 (0:00:01.284)       0:08:00.294 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=1)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=4)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=5)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=3)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=2)\u001b[0m\n\nTASK [ceph-osd : Systemd start osd] ********************************************\nSaturday 21 March 2026  00:52:41 +0000 (0:00:02.153)       0:08:02.448 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=1)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=4)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=3)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=5)\u001b[0m\n\nTASK [ceph-osd : Unset noup flag] **********************************************\nSaturday 21 March 2026  00:52:44 +0000 (0:00:03.861)       0:08:06.309 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-osd : Wait for all osd to be up] ************************************\nSaturday 21 March 2026  00:52:47 +0000 (0:00:03.044)       0:08:09.354 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-5 -> testbed-node-0]: Wait for all osd to be up (60 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-osd : Include crush_rules.yml] **************************************\nSaturday 21 March 2026  00:53:00 +0000 (0:00:12.863)       0:08:22.217 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nSaturday 21 March 2026  00:53:01 +0000 (0:00:00.738)       0:08:22.956 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Osds handler] **********************************\nSaturday 21 March 2026  00:53:02 +0000 (0:00:00.475)       0:08:23.431 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set_fact trigger_restart] **********************\nSaturday 21 March 2026  00:53:02 +0000 (0:00:00.447)       0:08:23.879 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _osd_handler_called before restart] ********\nSaturday 21 March 2026  00:53:02 +0000 (0:00:00.354)       0:08:24.233 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Unset noup flag] *******************************\nSaturday 21 March 2026  00:53:03 +0000 (0:00:00.289)       0:08:24.523 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy osd restart script] ***********************\nSaturday 21 March 2026  00:53:03 +0000 (0:00:00.625)       0:08:25.148 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Get pool list] *********************************\nSaturday 21 March 2026  00:53:03 +0000 (0:00:00.259)       0:08:25.408 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Get balancer module status] ********************\nSaturday 21 March 2026  00:53:04 +0000 (0:00:00.206)       0:08:25.615 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set_fact pools_pgautoscaler_mode] **************\nSaturday 21 March 2026  00:53:04 +0000 (0:00:00.217)       0:08:25.832 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Disable balancer] ******************************\nSaturday 21 March 2026  00:53:04 +0000 (0:00:00.108)       0:08:25.941 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Disable pg autoscale on pools] *****************\nSaturday 21 March 2026  00:53:04 +0000 (0:00:00.200)       0:08:26.141 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph osds daemon(s)] *******************\nSaturday 21 March 2026  00:53:04 +0000 (0:00:00.191)       0:08:26.333 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _osd_handler_called after restart] *********\nSaturday 21 March 2026  00:53:05 +0000 (0:00:00.364)       0:08:26.698 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Re-enable pg autoscale on pools] ***************\nSaturday 21 March 2026  00:53:05 +0000 (0:00:00.299)       0:08:26.997 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Re-enable balancer] ****************************\nSaturday 21 March 2026  00:53:06 +0000 (0:00:00.660)       0:08:27.658 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nPLAY [Apply role ceph-crash] ***************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nSaturday 21 March 2026  00:53:06 +0000 (0:00:00.601)       0:08:28.259 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nSaturday 21 March 2026  00:53:07 +0000 (0:00:01.061)       0:08:29.321 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nSaturday 21 March 2026  00:53:09 +0000 (0:00:01.185)       0:08:30.507 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nSaturday 21 March 2026  00:53:10 +0000 (0:00:01.096)       0:08:31.604 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nSaturday 21 March 2026  00:53:10 +0000 (0:00:00.772)       0:08:32.377 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nSaturday 21 March 2026  00:53:11 +0000 (0:00:00.822)       0:08:33.199 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nSaturday 21 March 2026  00:53:12 +0000 (0:00:00.671)       0:08:33.871 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nSaturday 21 March 2026  00:53:13 +0000 (0:00:01.190)       0:08:35.062 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nSaturday 21 March 2026  00:53:14 +0000 (0:00:00.663)       0:08:35.725 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nSaturday 21 March 2026  00:53:14 +0000 (0:00:00.633)       0:08:36.359 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nSaturday 21 March 2026  00:53:16 +0000 (0:00:01.496)       0:08:37.855 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nSaturday 21 March 2026  00:53:17 +0000 (0:00:01.193)       0:08:39.048 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nSaturday 21 March 2026  00:53:18 +0000 (0:00:00.952)       0:08:40.000 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nSaturday 21 March 2026  00:53:19 +0000 (0:00:00.548)       0:08:40.549 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nSaturday 21 March 2026  00:53:19 +0000 (0:00:00.728)       0:08:41.277 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nSaturday 21 March 2026  00:53:20 +0000 (0:00:00.519)       0:08:41.797 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nSaturday 21 March 2026  00:53:21 +0000 (0:00:00.839)       0:08:42.637 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nSaturday 21 March 2026  00:53:21 +0000 (0:00:00.525)       0:08:43.162 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nSaturday 21 March 2026  00:53:22 +0000 (0:00:00.767)       0:08:43.930 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nSaturday 21 March 2026  00:53:23 +0000 (0:00:00.619)       0:08:44.549 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nSaturday 21 March 2026  00:53:23 +0000 (0:00:00.803)       0:08:45.352 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-crash : Create client.crash keyring] ********************************\nSaturday 21 March 2026  00:53:24 +0000 (0:00:01.017)       0:08:46.370 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-crash : Get keys from monitors] *************************************\nSaturday 21 March 2026  00:53:29 +0000 (0:00:04.055)       0:08:50.425 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-crash : Copy ceph key(s) if needed] *********************************\nSaturday 21 March 2026  00:53:31 +0000 (0:00:02.105)       0:08:52.531 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-crash : Create /var/lib/ceph/crash/posted] **************************\nSaturday 21 March 2026  00:53:33 +0000 (0:00:02.031)       0:08:54.562 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-crash : Include_tasks systemd.yml] **********************************\nSaturday 21 March 2026  00:53:34 +0000 (0:00:01.488)       0:08:56.051 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-crash/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-crash : Generate systemd unit file for ceph-crash container] ********\nSaturday 21 March 2026  00:53:35 +0000 (0:00:00.904)       0:08:56.955 ******** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-crash : Start the ceph-crash service] *******************************\nSaturday 21 March 2026  00:53:37 +0000 (0:00:01.767)       0:08:58.722 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Ceph crash handler] ****************************\nSaturday 21 March 2026  00:53:41 +0000 (0:00:03.948)       0:09:02.671 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_crash.yml for testbed-node-4, testbed-node-3, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _crash_handler_called before restart] ******\nSaturday 21 March 2026  00:53:42 +0000 (0:00:01.434)       0:09:04.106 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart the ceph-crash service] ****************\nSaturday 21 March 2026  00:53:43 +0000 (0:00:00.694)       0:09:04.800 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _crash_handler_called after restart] *******\nSaturday 21 March 2026  00:53:46 +0000 (0:00:03.137)       0:09:07.938 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role ceph-mds] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nSaturday 21 March 2026  00:53:47 +0000 (0:00:00.885)       0:09:08.823 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nSaturday 21 March 2026  00:53:48 +0000 (0:00:00.855)       0:09:09.678 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nSaturday 21 March 2026  00:53:48 +0000 (0:00:00.572)       0:09:10.251 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nSaturday 21 March 2026  00:53:49 +0000 (0:00:00.793)       0:09:11.045 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nSaturday 21 March 2026  00:53:50 +0000 (0:00:00.819)       0:09:11.865 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nSaturday 21 March 2026  00:53:51 +0000 (0:00:00.717)       0:09:12.582 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nSaturday 21 March 2026  00:53:51 +0000 (0:00:00.780)       0:09:13.362 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nSaturday 21 March 2026  00:53:52 +0000 (0:00:00.785)       0:09:14.148 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nSaturday 21 March 2026  00:53:53 +0000 (0:00:00.342)       0:09:14.491 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nSaturday 21 March 2026  00:53:53 +0000 (0:00:00.319)       0:09:14.810 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nSaturday 21 March 2026  00:53:54 +0000 (0:00:00.680)       0:09:15.491 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nSaturday 21 March 2026  00:53:55 +0000 (0:00:01.091)       0:09:16.583 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nSaturday 21 March 2026  00:53:55 +0000 (0:00:00.307)       0:09:16.891 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nSaturday 21 March 2026  00:53:55 +0000 (0:00:00.339)       0:09:17.230 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nSaturday 21 March 2026  00:53:56 +0000 (0:00:00.434)       0:09:17.665 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nSaturday 21 March 2026  00:53:57 +0000 (0:00:00.807)       0:09:18.472 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nSaturday 21 March 2026  00:53:57 +0000 (0:00:00.385)       0:09:18.857 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nSaturday 21 March 2026  00:53:57 +0000 (0:00:00.344)       0:09:19.202 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nSaturday 21 March 2026  00:53:58 +0000 (0:00:00.373)       0:09:19.575 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nSaturday 21 March 2026  00:53:58 +0000 (0:00:00.622)       0:09:20.198 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nSaturday 21 March 2026  00:53:59 +0000 (0:00:00.344)       0:09:20.542 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Include create_mds_filesystems.yml] ***************************\nSaturday 21 March 2026  00:53:59 +0000 (0:00:00.581)       0:09:21.123 ******** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/create_mds_filesystems.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-facts : Get current default crush rule details] *********************\nSaturday 21 March 2026  00:54:00 +0000 (0:00:00.629)       0:09:21.753 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-facts : Get current default crush rule name] ************************\nSaturday 21 March 2026  00:54:02 +0000 (0:00:02.036)       0:09:23.790 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'rule_id': 0, 'rule_name': 'replicated_rule', 'type': 1, 'steps': [{'op': 'take', 'item': -1, 'item_name': 'default'}, {'op': 'chooseleaf_firstn', 'num': 0, 'type': 'host'}, {'op': 'emit'}]}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-mds : Create filesystem pools] **************************************\nSaturday 21 March 2026  00:54:02 +0000 (0:00:00.224)       0:09:24.014 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'application': 'cephfs', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'cephfs_data', 'pg_num': 16, 'pgp_num': 16, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'application': 'cephfs', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'cephfs_metadata', 'pg_num': 16, 'pgp_num': 16, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\nTASK [ceph-mds : Create ceph filesystem] ***************************************\nSaturday 21 March 2026  00:54:11 +0000 (0:00:09.263)       0:09:33.277 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-mds : Include common.yml] *******************************************\nSaturday 21 March 2026  00:54:15 +0000 (0:00:03.277)       0:09:36.555 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-mds : Create bootstrap-mds and mds directories] *********************\nSaturday 21 March 2026  00:54:15 +0000 (0:00:00.793)       0:09:37.349 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mds/ceph-testbed-node-3)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mds/ceph-testbed-node-4)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mds/ceph-testbed-node-5)\u001b[0m\n\nTASK [ceph-mds : Get keys from monitors] ***************************************\nSaturday 21 March 2026  00:54:17 +0000 (0:00:01.274)       0:09:38.623 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-mds : Copy ceph key(s) if needed] ***********************************\nSaturday 21 March 2026  00:54:19 +0000 (0:00:02.350)       0:09:40.974 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Create mds keyring] *******************************************\nSaturday 21 March 2026  00:54:20 +0000 (0:00:01.046)       0:09:42.020 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Non_containerized.yml] ****************************************\nSaturday 21 March 2026  00:54:22 +0000 (0:00:02.284)       0:09:44.304 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Containerized.yml] ********************************************\nSaturday 21 March 2026  00:54:23 +0000 (0:00:00.257)       0:09:44.562 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/containerized.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-mds : Include_tasks systemd.yml] ************************************\nSaturday 21 March 2026  00:54:23 +0000 (0:00:00.484)       0:09:45.047 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-mds : Generate systemd unit file] ***********************************\nSaturday 21 March 2026  00:54:24 +0000 (0:00:00.634)       0:09:45.681 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Generate systemd ceph-mds target file] ************************\nSaturday 21 March 2026  00:54:25 +0000 (0:00:01.287)       0:09:46.969 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Enable ceph-mds.target] ***************************************\nSaturday 21 March 2026  00:54:26 +0000 (0:00:01.374)       0:09:48.343 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [ceph-mds : Systemd start mds container] **********************************\nSaturday 21 March 2026  00:54:28 +0000 (0:00:01.971)       0:09:50.315 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [ceph-mds : Wait for mds socket to exist] *********************************\nSaturday 21 March 2026  00:54:31 +0000 (0:00:02.452)       0:09:52.767 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nSaturday 21 March 2026  00:54:32 +0000 (0:00:01.101)       0:09:53.868 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mdss handler] **********************************\nSaturday 21 March 2026  00:54:33 +0000 (0:00:00.868)       0:09:54.736 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mdss.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mds_handler_called before restart] ********\nSaturday 21 March 2026  00:54:33 +0000 (0:00:00.474)       0:09:55.211 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mds restart script] ***********************\nSaturday 21 March 2026  00:54:34 +0000 (0:00:00.286)       0:09:55.497 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mds daemon(s)] ********************\nSaturday 21 March 2026  00:54:35 +0000 (0:00:01.416)       0:09:56.914 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mds_handler_called after restart] *********\nSaturday 21 March 2026  00:54:36 +0000 (0:00:00.676)       0:09:57.591 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Apply role ceph-rgw] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nSaturday 21 March 2026  00:54:36 +0000 (0:00:00.534)       0:09:58.125 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nSaturday 21 March 2026  00:54:37 +0000 (0:00:00.701)       0:09:58.827 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nSaturday 21 March 2026  00:54:37 +0000 (0:00:00.483)       0:09:59.311 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nSaturday 21 March 2026  00:54:38 +0000 (0:00:00.420)       0:09:59.731 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nSaturday 21 March 2026  00:54:39 +0000 (0:00:00.710)       0:10:00.441 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nSaturday 21 March 2026  00:54:39 +0000 (0:00:00.734)       0:10:01.176 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nSaturday 21 March 2026  00:54:40 +0000 (0:00:00.746)       0:10:01.922 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nSaturday 21 March 2026  00:54:41 +0000 (0:00:00.601)       0:10:02.524 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nSaturday 21 March 2026  00:54:41 +0000 (0:00:00.306)       0:10:02.830 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nSaturday 21 March 2026  00:54:41 +0000 (0:00:00.292)       0:10:03.123 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nSaturday 21 March 2026  00:54:42 +0000 (0:00:00.668)       0:10:03.791 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nSaturday 21 March 2026  00:54:43 +0000 (0:00:00.978)       0:10:04.769 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nSaturday 21 March 2026  00:54:43 +0000 (0:00:00.364)       0:10:05.133 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nSaturday 21 March 2026  00:54:44 +0000 (0:00:00.337)       0:10:05.471 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nSaturday 21 March 2026  00:54:44 +0000 (0:00:00.355)       0:10:05.826 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nSaturday 21 March 2026  00:54:44 +0000 (0:00:00.573)       0:10:06.400 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nSaturday 21 March 2026  00:54:45 +0000 (0:00:00.339)       0:10:06.740 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nSaturday 21 March 2026  00:54:45 +0000 (0:00:00.360)       0:10:07.100 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nSaturday 21 March 2026  00:54:46 +0000 (0:00:00.367)       0:10:07.468 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nSaturday 21 March 2026  00:54:46 +0000 (0:00:00.634)       0:10:08.102 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nSaturday 21 March 2026  00:54:47 +0000 (0:00:00.370)       0:10:08.473 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Include common.yml] *******************************************\nSaturday 21 March 2026  00:54:47 +0000 (0:00:00.527)       0:10:09.000 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Get keys from monitors] ***************************************\nSaturday 21 March 2026  00:54:48 +0000 (0:00:00.921)       0:10:09.921 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************\nSaturday 21 March 2026  00:54:50 +0000 (0:00:02.124)       0:10:12.046 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Copy SSL certificate & key data to certificate path] **********\nSaturday 21 March 2026  00:54:51 +0000 (0:00:01.082)       0:10:13.129 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Include_tasks pre_requisite.yml] ******************************\nSaturday 21 March 2026  00:54:52 +0000 (0:00:00.333)       0:10:13.463 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/pre_requisite.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Create rados gateway directories] *****************************\nSaturday 21 March 2026  00:54:52 +0000 (0:00:00.801)       0:10:14.264 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-rgw : Create rgw keyrings] ******************************************\nSaturday 21 March 2026  00:54:53 +0000 (0:00:00.787)       0:10:15.051 ******** \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m\n\nTASK [ceph-rgw : Get keys from monitors] ***************************************\nSaturday 21 March 2026  00:54:57 +0000 (0:00:04.085)       0:10:19.136 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************\nSaturday 21 March 2026  00:55:00 +0000 (0:00:02.352)       0:10:21.488 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Rgw pool creation tasks] **************************************\nSaturday 21 March 2026  00:55:01 +0000 (0:00:01.111)       0:10:22.600 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/rgw_create_pools.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-rgw : Create ec profile] ********************************************\nSaturday 21 March 2026  00:55:01 +0000 (0:00:00.232)       0:10:22.832 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-rgw : Set crush rule] ***********************************************\nSaturday 21 March 2026  00:55:01 +0000 (0:00:00.566)       0:10:23.398 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-rgw : Create rgw pools] *********************************************\nSaturday 21 March 2026  00:55:02 +0000 (0:00:00.616)       0:10:24.015 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\nTASK [ceph-rgw : Include_tasks openstack-keystone.yml] *************************\nSaturday 21 March 2026  00:55:32 +0000 (0:00:29.874)       0:10:53.890 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Include_tasks start_radosgw.yml] ******************************\nSaturday 21 March 2026  00:55:32 +0000 (0:00:00.326)       0:10:54.216 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Include start_docker_rgw.yml] *********************************\nSaturday 21 March 2026  00:55:33 +0000 (0:00:00.637)       0:10:54.854 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/start_docker_rgw.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Include_task systemd.yml] *************************************\nSaturday 21 March 2026  00:55:33 +0000 (0:00:00.553)       0:10:55.407 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Generate systemd unit file] ***********************************\nSaturday 21 March 2026  00:55:34 +0000 (0:00:00.774)       0:10:56.182 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Generate systemd ceph-radosgw target file] ********************\nSaturday 21 March 2026  00:55:35 +0000 (0:00:01.148)       0:10:57.330 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Enable ceph-radosgw.target] ***********************************\nSaturday 21 March 2026  00:55:36 +0000 (0:00:01.078)       0:10:58.409 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [ceph-rgw : Systemd start rgw container] **********************************\nSaturday 21 March 2026  00:55:38 +0000 (0:00:01.663)       0:11:00.072 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nSaturday 21 March 2026  00:55:41 +0000 (0:00:02.415)       0:11:02.488 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rgws handler] **********************************\nSaturday 21 March 2026  00:55:41 +0000 (0:00:00.305)       0:11:02.794 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_rgws.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _rgw_handler_called before restart] ********\nSaturday 21 March 2026  00:55:42 +0000 (0:00:00.805)       0:11:03.599 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy rgw restart script] ***********************\nSaturday 21 March 2026  00:55:42 +0000 (0:00:00.315)       0:11:03.915 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph rgw daemon(s)] ********************\nSaturday 21 March 2026  00:55:42 +0000 (0:00:00.308)       0:11:04.223 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _rgw_handler_called after restart] *********\nSaturday 21 March 2026  00:55:43 +0000 (0:00:01.116)       0:11:05.339 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=134 \u001b[0m \u001b[0;33mchanged=35  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=125 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=127 \u001b[0m \u001b[0;33mchanged=31  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=120 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=134 \u001b[0m \u001b[0;33mchanged=33  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=119 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=193 \u001b[0m \u001b[0;33mchanged=45  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=162 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=40  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=123 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=41  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=121 \u001b[0m rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:55:44 +0000 (0:00:00.245)       0:11:05.585 ******** \n=============================================================================== \nceph-container-common : Pulling Ceph container image ------------------- 58.61s\nceph-osd : Use ceph-volume to create osds ------------------------------ 44.17s\nceph-mgr : Wait for all mgr to be up ----------------------------------- 36.71s\nceph-rgw : Create rgw pools -------------------------------------------- 29.88s\nceph-mon : Waiting for the monitor(s) to form the quorum... ------------ 21.52s\nceph-mon : Set cluster configs ----------------------------------------- 14.96s\nceph-osd : Wait for all osd to be up ----------------------------------- 12.86s\nceph-mgr : Create ceph mgr keyring(s) on a mon node -------------------- 10.78s\nceph-mds : Create filesystem pools -------------------------------------- 9.26s\nceph-mon : Fetch ceph initial keys -------------------------------------- 8.91s\nceph-config : Create ceph initial directories --------------------------- 6.69s\nceph-mgr : Disable ceph mgr enabled modules ----------------------------- 6.39s\nceph-mgr : Add modules to ceph-mgr -------------------------------------- 4.81s\nceph-rgw : Create rgw keyrings ------------------------------------------ 4.09s\nceph-crash : Create client.crash keyring -------------------------------- 4.06s\nceph-mon : Copy admin keyring over to mons ------------------------------ 4.01s\nceph-osd : Apply operating system tuning -------------------------------- 3.96s\nceph-crash : Start the ceph-crash service ------------------------------- 3.95s\nceph-osd : Systemd start osd -------------------------------------------- 3.86s\nceph-facts : Set_fact _monitor_addresses - ipv4 ------------------------- 3.59s\n\u001b[32m2026-03-21 00:55:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:55:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:55:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:55:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:55:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:55:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:56:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state SUCCESS\u001b[0m\n\n\nPLAY [Set kolla_action_mariadb] ************************************************\n\nTASK [Inform the user about the following task] ********************************\nSaturday 21 March 2026  00:53:10 +0000 (0:00:00.092)       0:00:00.092 ******** \n\u001b[0;32mok: [localhost] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"The task 'Check MariaDB service' fails if the MariaDB service has not yet been deployed. This is fine.\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Check MariaDB service] ***************************************************\nSaturday 21 March 2026  00:53:11 +0000 (0:00:00.035)       0:00:00.128 ******** \n\u001b[0;31mfatal: [localhost]: FAILED! => {\"changed\": false, \"elapsed\": 2, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.9:3306\"}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\nTASK [Set kolla_action_mariadb = upgrade if MariaDB is already running] ********\nSaturday 21 March 2026  00:53:13 +0000 (0:00:02.814)       0:00:02.942 ******** \n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nTASK [Set kolla_action_mariadb = kolla_action_ng] ******************************\nSaturday 21 March 2026  00:53:13 +0000 (0:00:00.052)       0:00:02.995 ******** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  00:53:14 +0000 (0:00:00.292)       0:00:03.287 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  00:53:14 +0000 (0:00:00.305)       0:00:03.593 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_mariadb_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_mariadb_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_mariadb_True)\u001b[0m\n\nPLAY [Apply role mariadb] ******************************************************\n\nTASK [mariadb : Group MariaDB hosts based on shards] ***************************\nSaturday 21 March 2026  00:53:15 +0000 (0:00:00.566)       0:00:04.159 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nSaturday 21 March 2026  00:53:15 +0000 (0:00:00.501)       0:00:04.661 ******** \n\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [mariadb : Ensuring config directories exist] *****************************\nSaturday 21 March 2026  00:53:16 +0000 (0:00:00.965)       0:00:05.626 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\nTASK [mariadb : Ensuring database backup config directory exists] **************\nSaturday 21 March 2026  00:53:19 +0000 (0:00:03.311)       0:00:08.938 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Copying over my.cnf for mariabackup] ***************************\nSaturday 21 March 2026  00:53:20 +0000 (0:00:00.756)       0:00:09.694 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Copying over config.json files for services] *******************\nSaturday 21 March 2026  00:53:21 +0000 (0:00:01.415)       0:00:11.110 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\nTASK [mariadb : Copying over config.json files for mariabackup] ****************\nSaturday 21 March 2026  00:53:26 +0000 (0:00:04.193)       0:00:15.303 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Copying over galera.cnf] ***************************************\nSaturday 21 March 2026  00:53:27 +0000 (0:00:01.093)       0:00:16.397 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nSaturday 21 March 2026  00:53:30 +0000 (0:00:03.376)       0:00:19.774 ******** \n\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over extra CA certificates] ********\nSaturday 21 March 2026  00:53:31 +0000 (0:00:00.488)       0:00:20.263 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] ***\nSaturday 21 March 2026  00:53:34 +0000 (0:00:03.419)       0:00:23.682 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over backend internal TLS key] *****\nSaturday 21 March 2026  00:53:36 +0000 (0:00:02.288)       0:00:25.970 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Check mariadb containers] **************************************\nSaturday 21 March 2026  00:53:39 +0000 (0:00:02.711)       0:00:28.682 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\nTASK [mariadb : Create MariaDB volume] *****************************************\nSaturday 21 March 2026  00:53:43 +0000 (0:00:03.503)       0:00:32.185 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Divide hosts by their MariaDB volume availability] *************\nSaturday 21 March 2026  00:53:43 +0000 (0:00:00.811)       0:00:32.997 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Establish whether the cluster has already existed] *************\nSaturday 21 March 2026  00:53:44 +0000 (0:00:00.349)       0:00:33.347 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Check MariaDB service port liveness] ***************************\nSaturday 21 March 2026  00:53:44 +0000 (0:00:00.328)       0:00:33.675 ******** \n\u001b[0;31mfatal: [testbed-node-0]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.10:3306\"}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\u001b[0;31mfatal: [testbed-node-1]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.11:3306\"}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\u001b[0;31mfatal: [testbed-node-2]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.12:3306\"}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\nTASK [mariadb : Divide hosts by their MariaDB service port liveness] ***********\nSaturday 21 March 2026  00:53:55 +0000 (0:00:11.221)       0:00:44.896 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Fail on existing but stopped cluster] **************************\nSaturday 21 March 2026  00:53:56 +0000 (0:00:00.492)       0:00:45.389 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Check MariaDB service WSREP sync status] ***********************\nSaturday 21 March 2026  00:53:56 +0000 (0:00:00.440)       0:00:45.830 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Extract MariaDB service WSREP sync status] *********************\nSaturday 21 March 2026  00:53:57 +0000 (0:00:00.515)       0:00:46.345 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Divide hosts by their MariaDB service WSREP sync status] *******\nSaturday 21 March 2026  00:53:57 +0000 (0:00:00.694)       0:00:47.039 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Fail when MariaDB services are not synced across the whole cluster] ***\nSaturday 21 March 2026  00:53:58 +0000 (0:00:00.486)       0:00:47.525 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nSaturday 21 March 2026  00:53:58 +0000 (0:00:00.457)       0:00:47.983 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/bootstrap_cluster.yml for testbed-node-0\u001b[0m\n\nTASK [mariadb : Running MariaDB bootstrap container] ***************************\nSaturday 21 March 2026  00:53:59 +0000 (0:00:00.398)       0:00:48.382 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Store bootstrap host name into facts] **************************\nSaturday 21 March 2026  00:54:09 +0000 (0:00:10.698)       0:00:59.080 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nSaturday 21 March 2026  00:54:10 +0000 (0:00:00.274)       0:00:59.355 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [mariadb : Starting first MariaDB container] *******************\nSaturday 21 March 2026  00:54:11 +0000 (0:00:00.859)       0:01:00.214 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [mariadb : Wait for first MariaDB service port liveness] *******\nSaturday 21 March 2026  00:54:18 +0000 (0:00:07.587)       0:01:07.802 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [mariadb : Wait for first MariaDB service to sync WSREP] *******\nSaturday 21 March 2026  00:54:20 +0000 (0:00:01.528)       0:01:09.331 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [mariadb : Ensure MariaDB is running normally on bootstrap host] ***\nSaturday 21 March 2026  00:54:22 +0000 (0:00:01.901)       0:01:11.232 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [mariadb : Restart MariaDB on existing cluster members] ********\nSaturday 21 March 2026  00:54:22 +0000 (0:00:00.220)       0:01:11.453 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [mariadb : Start MariaDB on new nodes] *************************\nSaturday 21 March 2026  00:54:22 +0000 (0:00:00.297)       0:01:11.751 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_restart\u001b[0m\n\nPLAY [Restart mariadb services] ************************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Start mariadb services] **************************************************\n\nTASK [mariadb : Restart MariaDB container] *************************************\nSaturday 21 March 2026  00:54:22 +0000 (0:00:00.276)       0:01:12.027 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service port liveness] ************************\nSaturday 21 March 2026  00:54:44 +0000 (0:00:21.805)       0:01:33.833 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to sync WSREP] ************************\nSaturday 21 March 2026  00:54:54 +0000 (0:00:09.593)       0:01:43.426 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nPLAY [Start mariadb services] **************************************************\n\nTASK [mariadb : Restart MariaDB container] *************************************\nSaturday 21 March 2026  00:54:56 +0000 (0:00:02.447)       0:01:45.874 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service port liveness] ************************\nSaturday 21 March 2026  00:55:13 +0000 (0:00:17.000)       0:02:02.875 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to sync WSREP] ************************\nSaturday 21 March 2026  00:55:28 +0000 (0:00:14.938)       0:02:17.813 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Restart bootstrap mariadb service] ***************************************\n\nTASK [mariadb : Restart MariaDB container] *************************************\nSaturday 21 March 2026  00:55:31 +0000 (0:00:02.651)       0:02:20.465 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service port liveness] ************************\nSaturday 21 March 2026  00:55:42 +0000 (0:00:11.109)       0:02:31.574 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to sync WSREP] ************************\nSaturday 21 March 2026  00:55:45 +0000 (0:00:03.486)       0:02:35.061 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nPLAY [Apply mariadb post-configuration] ****************************************\n\nTASK [Include mariadb post-deploy.yml] *****************************************\nSaturday 21 March 2026  00:55:48 +0000 (0:00:02.445)       0:02:37.506 ******** \n\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [mariadb : Creating shard root mysql user] ********************************\nSaturday 21 March 2026  00:55:49 +0000 (0:00:00.662)       0:02:38.168 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Creating mysql monitor user] ***********************************\nSaturday 21 March 2026  00:55:51 +0000 (0:00:02.281)       0:02:40.450 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Creating database backup user and setting permissions] *********\nSaturday 21 March 2026  00:55:53 +0000 (0:00:01.902)       0:02:42.352 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Granting permissions on Mariabackup database to backup user] ***\nSaturday 21 March 2026  00:55:55 +0000 (0:00:01.849)       0:02:44.202 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to be ready through VIP] **************\nSaturday 21 March 2026  00:55:57 +0000 (0:00:02.324)       0:02:46.526 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Include mariadb post-upgrade.yml] ****************************************\nSaturday 21 March 2026  00:56:00 +0000 (0:00:02.833)       0:02:49.360 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mlocalhost\u001b[0m                  : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=18  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=18  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:56:00 +0000 (0:00:00.244)       0:02:49.604 ******** \n=============================================================================== \nmariadb : Restart MariaDB container ------------------------------------ 38.81s\nmariadb : Wait for MariaDB service port liveness ----------------------- 24.53s\nmariadb : Check MariaDB service port liveness -------------------------- 11.22s\nmariadb : Restart MariaDB container ------------------------------------ 11.11s\nmariadb : Running MariaDB bootstrap container -------------------------- 10.70s\nmariadb : Starting first MariaDB container ------------------------------ 7.59s\nmariadb : Wait for MariaDB service to sync WSREP ------------------------ 5.10s\nmariadb : Copying over config.json files for services ------------------- 4.19s\nmariadb : Check mariadb containers -------------------------------------- 3.50s\nmariadb : Wait for MariaDB service port liveness ------------------------ 3.49s\nservice-cert-copy : mariadb | Copying over extra CA certificates -------- 3.42s\nmariadb : Copying over galera.cnf --------------------------------------- 3.38s\nmariadb : Ensuring config directories exist ----------------------------- 3.31s\nmariadb : Wait for MariaDB service to be ready through VIP -------------- 2.83s\nCheck MariaDB service --------------------------------------------------- 2.81s\nservice-cert-copy : mariadb | Copying over backend internal TLS key ----- 2.71s\nmariadb : Wait for MariaDB service to sync WSREP ------------------------ 2.45s\nmariadb : Granting permissions on Mariabackup database to backup user --- 2.32s\nservice-cert-copy : mariadb | Copying over backend internal TLS certificate --- 2.29s\nmariadb : Creating shard root mysql user -------------------------------- 2.28s\n\u001b[32m2026-03-21 00:56:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:56:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:56:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:56:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:56:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:56:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:56:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:56:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:56:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:56:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:56:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:56:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:56:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:56:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:56:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:56:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:56:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:56:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:56:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:56:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:57:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:57:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:57:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:57:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:57:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:57:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:57:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:57:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:57:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:57:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:57:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:57:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:57:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:57:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  00:56:03 +0000 (0:00:00.282)       0:00:00.282 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  00:56:04 +0000 (0:00:00.243)       0:00:00.525 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_horizon_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_horizon_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_horizon_True)\u001b[0m\n\nPLAY [Apply role horizon] ******************************************************\n\nTASK [horizon : include_tasks] *************************************************\nSaturday 21 March 2026  00:56:04 +0000 (0:00:00.279)       0:00:00.805 ******** \n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [horizon : Ensuring config directories exist] *****************************\nSaturday 21 March 2026  00:56:04 +0000 (0:00:00.441)       0:00:01.247 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\nTASK [horizon : Set empty custom policy] ***************************************\nSaturday 21 March 2026  00:56:06 +0000 (0:00:01.774)       0:00:03.021 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : include_tasks] *************************************************\nSaturday 21 March 2026  00:56:06 +0000 (0:00:00.260)       0:00:03.282 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'trove', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'trove', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'trove', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'ceilometer', 'enabled': 'yes'})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'cinder', 'enabled': 'yes'})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'designate', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'glance', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'keystone', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'magnum', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'manila', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'neutron', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'nova', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'octavia', 'enabled': True})\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nSaturday 21 March 2026  00:56:07 +0000 (0:00:00.630)       0:00:03.912 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nSaturday 21 March 2026  00:56:07 +0000 (0:00:00.370)       0:00:04.283 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nSaturday 21 March 2026  00:56:07 +0000 (0:00:00.140)       0:00:04.423 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nSaturday 21 March 2026  00:56:08 +0000 (0:00:00.238)       0:00:04.662 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nSaturday 21 March 2026  00:56:08 +0000 (0:00:00.273)       0:00:04.936 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nSaturday 21 March 2026  00:56:08 +0000 (0:00:00.115)       0:00:05.052 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nSaturday 21 March 2026  00:56:08 +0000 (0:00:00.365)       0:00:05.417 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nSaturday 21 March 2026  00:56:09 +0000 (0:00:00.387)       0:00:05.804 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nSaturday 21 March 2026  00:56:09 +0000 (0:00:00.179)       0:00:05.984 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nSaturday 21 March 2026  00:56:09 +0000 (0:00:00.329)       0:00:06.313 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nSaturday 21 March 2026  00:56:10 +0000 (0:00:00.260)       0:00:06.574 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nSaturday 21 March 2026  00:56:10 +0000 (0:00:00.118)       0:00:06.693 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nSaturday 21 March 2026  00:56:10 +0000 (0:00:00.424)       0:00:07.118 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nSaturday 21 March 2026  00:56:10 +0000 (0:00:00.283)       0:00:07.401 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nSaturday 21 March 2026  00:56:11 +0000 (0:00:00.101)       0:00:07.503 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nSaturday 21 March 2026  00:56:11 +0000 (0:00:00.251)       0:00:07.755 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nSaturday 21 March 2026  00:56:11 +0000 (0:00:00.395)       0:00:08.151 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nSaturday 21 March 2026  00:56:11 +0000 (0:00:00.101)       0:00:08.253 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nSaturday 21 March 2026  00:56:12 +0000 (0:00:00.256)       0:00:08.509 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nSaturday 21 March 2026  00:56:12 +0000 (0:00:00.366)       0:00:08.876 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nSaturday 21 March 2026  00:56:12 +0000 (0:00:00.089)       0:00:08.966 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nSaturday 21 March 2026  00:56:12 +0000 (0:00:00.250)       0:00:09.217 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nSaturday 21 March 2026  00:56:13 +0000 (0:00:00.416)       0:00:09.633 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nSaturday 21 March 2026  00:56:13 +0000 (0:00:00.102)       0:00:09.736 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nSaturday 21 March 2026  00:56:13 +0000 (0:00:00.253)       0:00:09.989 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nSaturday 21 March 2026  00:56:14 +0000 (0:00:00.484)       0:00:10.474 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nSaturday 21 March 2026  00:56:14 +0000 (0:00:00.139)       0:00:10.614 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nSaturday 21 March 2026  00:56:14 +0000 (0:00:00.269)       0:00:10.883 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nSaturday 21 March 2026  00:56:14 +0000 (0:00:00.480)       0:00:11.364 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nSaturday 21 March 2026  00:56:15 +0000 (0:00:00.126)       0:00:11.491 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Copying over config.json files for services] *******************\nSaturday 21 March 2026  00:56:15 +0000 (0:00:00.289)       0:00:11.781 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Copying over horizon.conf] *************************************\nSaturday 21 March 2026  00:56:17 +0000 (0:00:01.903)       0:00:13.684 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m\n\nTASK [horizon : Copying over kolla-settings.py] ********************************\nSaturday 21 March 2026  00:56:19 +0000 (0:00:02.765)       0:00:16.450 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m\n\nTASK [horizon : Copying over custom-settings.py] *******************************\nSaturday 21 March 2026  00:56:21 +0000 (0:00:01.935)       0:00:18.386 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m\n\nTASK [horizon : Copying over existing policy file] *****************************\nSaturday 21 March 2026  00:56:23 +0000 (0:00:01.749)       0:00:20.135 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Copying over custom themes] ************************************\nSaturday 21 March 2026  00:56:23 +0000 (0:00:00.282)       0:00:20.417 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : include_tasks] *************************************************\nSaturday 21 March 2026  00:56:24 +0000 (0:00:00.277)       0:00:20.695 ******** \n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : horizon | Copying over extra CA certificates] ********\nSaturday 21 March 2026  00:56:25 +0000 (0:00:00.844)       0:00:21.540 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\nTASK [service-cert-copy : horizon | Copying over backend internal TLS certificate] ***\nSaturday 21 March 2026  00:56:26 +0000 (0:00:01.593)       0:00:23.133 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : horizon | Copying over backend internal TLS key] *****\nSaturday 21 March 2026  00:56:27 +0000 (0:00:00.927)       0:00:24.060 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Deploy horizon container] **************************************\nSaturday 21 March 2026  00:56:28 +0000 (0:00:01.000)       0:00:25.060 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\nTASK [horizon : include_tasks] *************************************************\nSaturday 21 March 2026  00:56:29 +0000 (0:00:01.381)       0:00:26.442 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : include_tasks] *************************************************\nSaturday 21 March 2026  00:56:30 +0000 (0:00:00.307)       0:00:26.749 ******** \n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/bootstrap.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [horizon : Creating Horizon database] *************************************\nSaturday 21 March 2026  00:56:31 +0000 (0:00:00.724)       0:00:27.474 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Creating Horizon database user and setting permissions] ********\nSaturday 21 March 2026  00:56:33 +0000 (0:00:02.693)       0:00:30.168 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Running Horizon bootstrap container] ***************************\nSaturday 21 March 2026  00:56:35 +0000 (0:00:02.052)       0:00:32.220 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Flush handlers] ************************************************\nSaturday 21 March 2026  00:56:50 +0000 (0:00:15.090)       0:00:47.310 ******** \n\nTASK [horizon : Flush handlers] ************************************************\nSaturday 21 March 2026  00:56:50 +0000 (0:00:00.065)       0:00:47.376 ******** \n\nTASK [horizon : Flush handlers] ************************************************\nSaturday 21 March 2026  00:56:50 +0000 (0:00:00.081)       0:00:47.457 ******** \n\nRUNNING HANDLER [horizon : Restart horizon container] **************************\nSaturday 21 March 2026  00:56:51 +0000 (0:00:00.088)       0:00:47.546 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=37  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=15  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=15  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:57:36 +0000 (0:00:45.880)       0:01:33.427 ******** \n=============================================================================== \nhorizon : Restart horizon container ------------------------------------ 45.88s\nhorizon : Running Horizon bootstrap container -------------------------- 15.09s\nhorizon : Copying over horizon.conf ------------------------------------- 2.77s\nhorizon : Creating Horizon database ------------------------------------- 2.69s\nhorizon : Creating Horizon database user and setting permissions -------- 2.05s\nhorizon : Copying over kolla-settings.py -------------------------------- 1.94s\nhorizon : Copying over config.json files for services ------------------- 1.90s\nhorizon : Ensuring config directories exist ----------------------------- 1.77s\nhorizon : Copying over custom-settings.py ------------------------------- 1.75s\nservice-cert-copy : horizon | Copying over extra CA certificates -------- 1.59s\nhorizon : Deploy horizon container -------------------------------------- 1.38s\nservice-cert-copy : horizon | Copying over backend internal TLS key ----- 1.00s\nservice-cert-copy : horizon | Copying over backend internal TLS certificate --- 0.93s\nhorizon : include_tasks ------------------------------------------------- 0.84s\nhorizon : include_tasks ------------------------------------------------- 0.72s\nhorizon : include_tasks ------------------------------------------------- 0.63s\nhorizon : Update policy file name --------------------------------------- 0.48s\nhorizon : Update policy file name --------------------------------------- 0.48s\nhorizon : include_tasks ------------------------------------------------- 0.44s\nhorizon : Update custom policy file name -------------------------------- 0.42s\n\u001b[32m2026-03-21 00:57:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:57:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:57:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:57:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:57:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:57:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state SUCCESS\u001b[0m\n\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Create ceph pools] *******************************************************\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nSaturday 21 March 2026  00:55:49 +0000 (0:00:00.551)       0:00:00.551 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nSaturday 21 March 2026  00:55:49 +0000 (0:00:00.474)       0:00:01.025 ******** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nSaturday 21 March 2026  00:55:50 +0000 (0:00:00.844)       0:00:01.869 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nSaturday 21 March 2026  00:55:50 +0000 (0:00:00.251)       0:00:02.121 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nSaturday 21 March 2026  00:55:51 +0000 (0:00:00.681)       0:00:02.802 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nSaturday 21 March 2026  00:55:51 +0000 (0:00:00.291)       0:00:03.094 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nSaturday 21 March 2026  00:55:52 +0000 (0:00:00.260)       0:00:03.355 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nSaturday 21 March 2026  00:55:52 +0000 (0:00:00.252)       0:00:03.608 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nSaturday 21 March 2026  00:55:52 +0000 (0:00:00.391)       0:00:03.999 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nSaturday 21 March 2026  00:55:53 +0000 (0:00:00.286)       0:00:04.286 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nSaturday 21 March 2026  00:55:53 +0000 (0:00:00.635)       0:00:04.922 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nSaturday 21 March 2026  00:55:54 +0000 (0:00:00.407)       0:00:05.329 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nSaturday 21 March 2026  00:55:56 +0000 (0:00:02.786)       0:00:08.115 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nSaturday 21 March 2026  00:55:57 +0000 (0:00:00.419)       0:00:08.535 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nSaturday 21 March 2026  00:55:58 +0000 (0:00:00.838)       0:00:09.374 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nSaturday 21 March 2026  00:55:58 +0000 (0:00:00.186)       0:00:09.560 ******** \n\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': '22929cdeac3a', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-21 00:55:54.962395', 'end': '2026-03-21 00:55:54.995637', 'delta': '0:00:00.033242', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['22929cdeac3a'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': 'a96b2c5299e5', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-21 00:55:55.943631', 'end': '2026-03-21 00:55:55.973282', 'delta': '0:00:00.029651', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['a96b2c5299e5'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': '57e0bd4375ae', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-21 00:55:56.696837', 'end': '2026-03-21 00:55:56.725035', 'delta': '0:00:00.028198', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['57e0bd4375ae'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nSaturday 21 March 2026  00:55:58 +0000 (0:00:00.376)       0:00:09.937 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nSaturday 21 March 2026  00:55:59 +0000 (0:00:00.554)       0:00:10.491 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nSaturday 21 March 2026  00:56:00 +0000 (0:00:01.599)       0:00:12.091 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nSaturday 21 March 2026  00:56:01 +0000 (0:00:00.274)       0:00:12.365 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nSaturday 21 March 2026  00:56:01 +0000 (0:00:00.433)       0:00:12.798 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nSaturday 21 March 2026  00:56:02 +0000 (0:00:00.477)       0:00:13.276 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nSaturday 21 March 2026  00:56:02 +0000 (0:00:00.121)       0:00:13.398 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nSaturday 21 March 2026  00:56:02 +0000 (0:00:00.205)       0:00:13.604 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nSaturday 21 March 2026  00:56:02 +0000 (0:00:00.238)       0:00:13.842 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nSaturday 21 March 2026  00:56:02 +0000 (0:00:00.285)       0:00:14.127 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nSaturday 21 March 2026  00:56:03 +0000 (0:00:00.403)       0:00:14.530 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nSaturday 21 March 2026  00:56:03 +0000 (0:00:00.273)       0:00:14.804 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nSaturday 21 March 2026  00:56:03 +0000 (0:00:00.269)       0:00:15.073 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nSaturday 21 March 2026  00:56:04 +0000 (0:00:00.263)       0:00:15.337 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nSaturday 21 March 2026  00:56:04 +0000 (0:00:00.394)       0:00:15.731 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--64dd100a--6bcc--5763--b557--0732c79bfd55-osd--block--64dd100a--6bcc--5763--b557--0732c79bfd55', 'dm-uuid-LVM-MWn87OlXxZ0hjMuGG89puNwiFbMviFcm1xAms0qPz8LYU8Ut0AGzeF9WoI61D5NM'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--5368e13a--d21f--5912--91bd--da537815d9e8-osd--block--5368e13a--d21f--5912--91bd--da537815d9e8', 'dm-uuid-LVM-xgLPNxLNUSwUnkWnyYv6gd71t6QIH1sH9rPUXF3date4d4aBKBICcuqoW7mNU6TI'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--ed04aa0b--4f5a--523b--a5d2--05072a1b260d-osd--block--ed04aa0b--4f5a--523b--a5d2--05072a1b260d', 'dm-uuid-LVM-viMZkgfP3jZUgTAo0bSRpd40fLtFkJJmO8XW5mmi1aRChVrq8X3DDkwGl8XRb5at'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part1', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part14', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part15', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part16', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--45b09425--beda--50b4--a83b--1fc50b5adca4-osd--block--45b09425--beda--50b4--a83b--1fc50b5adca4', 'dm-uuid-LVM-XzfTGfmLHoftdDBti3ot8l7sbRObO8yALj7bfZRysK7QQ9WjzIgwF2E0nO1Pc63W'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'holders': ['ceph--64dd100a--6bcc--5763--b557--0732c79bfd55-osd--block--64dd100a--6bcc--5763--b557--0732c79bfd55'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-HiPrG3-9qzo-WJZ0-1Smx-Bi1t-ReBz-6cdxta', 'scsi-0QEMU_QEMU_HARDDISK_f08311eb-e2ad-4eb7-8da9-ac6d899a2409', 'scsi-SQEMU_QEMU_HARDDISK_f08311eb-e2ad-4eb7-8da9-ac6d899a2409'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'holders': ['ceph--5368e13a--d21f--5912--91bd--da537815d9e8-osd--block--5368e13a--d21f--5912--91bd--da537815d9e8'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-m0F2nF-j9UB-mZfT-bRv8-xW1c-EVch-ZfDvHt', 'scsi-0QEMU_QEMU_HARDDISK_e03b171b-b829-41a0-a0b6-cc46d85f013d', 'scsi-SQEMU_QEMU_HARDDISK_e03b171b-b829-41a0-a0b6-cc46d85f013d'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_9fa3c153-0e2a-45ef-8e48-e0095eb937ad', 'scsi-SQEMU_QEMU_HARDDISK_9fa3c153-0e2a-45ef-8e48-e0095eb937ad'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-21-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part1', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part14', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part15', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part16', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'holders': ['ceph--ed04aa0b--4f5a--523b--a5d2--05072a1b260d-osd--block--ed04aa0b--4f5a--523b--a5d2--05072a1b260d'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-nra7ru-O0w6-dM26-lz5o-2arZ-rXsE-lANagX', 'scsi-0QEMU_QEMU_HARDDISK_4e74a544-8b3a-4a05-98a1-b7cc13f13c18', 'scsi-SQEMU_QEMU_HARDDISK_4e74a544-8b3a-4a05-98a1-b7cc13f13c18'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'holders': ['ceph--45b09425--beda--50b4--a83b--1fc50b5adca4-osd--block--45b09425--beda--50b4--a83b--1fc50b5adca4'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-XQTe4H-QTjK-DLSs-Qjj6-RKjC-da4T-xXChra', 'scsi-0QEMU_QEMU_HARDDISK_666e73fe-87f5-4f65-9ff2-95a038c8b18d', 'scsi-SQEMU_QEMU_HARDDISK_666e73fe-87f5-4f65-9ff2-95a038c8b18d'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ae4aabb-0174-4999-a1d0-4bf78cbb2965', 'scsi-SQEMU_QEMU_HARDDISK_8ae4aabb-0174-4999-a1d0-4bf78cbb2965'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-14-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--ede169e1--d974--55ed--a938--c303cb89cb5b-osd--block--ede169e1--d974--55ed--a938--c303cb89cb5b', 'dm-uuid-LVM-bEBXWAqtQdfT6xKYbL0dF10We9CLdKgKh0pf0vpR2lQnL405kp4ndXs76m3o0J5Y'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a-osd--block--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a', 'dm-uuid-LVM-G1LwYmzPCvwLZRWX1y5oYSHlzZdIxlOhfbzzvktDUwfRIgHaiaw6r0MV8HVh1TLk'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part1', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part14', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part15', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part16', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'holders': ['ceph--ede169e1--d974--55ed--a938--c303cb89cb5b-osd--block--ede169e1--d974--55ed--a938--c303cb89cb5b'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-HVhEyD-n8kM-PvHm-CUDD-tuRZ-igjC-Nzour2', 'scsi-0QEMU_QEMU_HARDDISK_f76ae8f0-65cc-4a62-b19c-23f8a40fbde6', 'scsi-SQEMU_QEMU_HARDDISK_f76ae8f0-65cc-4a62-b19c-23f8a40fbde6'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'holders': ['ceph--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a-osd--block--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-YifJdB-JYWE-HhRV-mF1g-0k83-hzWG-ctJZWy', 'scsi-0QEMU_QEMU_HARDDISK_d7bb81ba-6491-4b5a-bb82-f0b49314eac0', 'scsi-SQEMU_QEMU_HARDDISK_d7bb81ba-6491-4b5a-bb82-f0b49314eac0'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_cdffe5f1-7f0d-460e-a1db-d6846d226144', 'scsi-SQEMU_QEMU_HARDDISK_cdffe5f1-7f0d-460e-a1db-d6846d226144'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-28-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nSaturday 21 March 2026  00:56:05 +0000 (0:00:00.551)       0:00:16.282 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--64dd100a--6bcc--5763--b557--0732c79bfd55-osd--block--64dd100a--6bcc--5763--b557--0732c79bfd55', 'dm-uuid-LVM-MWn87OlXxZ0hjMuGG89puNwiFbMviFcm1xAms0qPz8LYU8Ut0AGzeF9WoI61D5NM'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--5368e13a--d21f--5912--91bd--da537815d9e8-osd--block--5368e13a--d21f--5912--91bd--da537815d9e8', 'dm-uuid-LVM-xgLPNxLNUSwUnkWnyYv6gd71t6QIH1sH9rPUXF3date4d4aBKBICcuqoW7mNU6TI'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--ed04aa0b--4f5a--523b--a5d2--05072a1b260d-osd--block--ed04aa0b--4f5a--523b--a5d2--05072a1b260d', 'dm-uuid-LVM-viMZkgfP3jZUgTAo0bSRpd40fLtFkJJmO8XW5mmi1aRChVrq8X3DDkwGl8XRb5at'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--45b09425--beda--50b4--a83b--1fc50b5adca4-osd--block--45b09425--beda--50b4--a83b--1fc50b5adca4', 'dm-uuid-LVM-XzfTGfmLHoftdDBti3ot8l7sbRObO8yALj7bfZRysK7QQ9WjzIgwF2E0nO1Pc63W'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--ede169e1--d974--55ed--a938--c303cb89cb5b-osd--block--ede169e1--d974--55ed--a938--c303cb89cb5b', 'dm-uuid-LVM-bEBXWAqtQdfT6xKYbL0dF10We9CLdKgKh0pf0vpR2lQnL405kp4ndXs76m3o0J5Y'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a-osd--block--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a', 'dm-uuid-LVM-G1LwYmzPCvwLZRWX1y5oYSHlzZdIxlOhfbzzvktDUwfRIgHaiaw6r0MV8HVh1TLk'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part1', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part14', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part15', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part16', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--64dd100a--6bcc--5763--b557--0732c79bfd55-osd--block--64dd100a--6bcc--5763--b557--0732c79bfd55'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-HiPrG3-9qzo-WJZ0-1Smx-Bi1t-ReBz-6cdxta', 'scsi-0QEMU_QEMU_HARDDISK_f08311eb-e2ad-4eb7-8da9-ac6d899a2409', 'scsi-SQEMU_QEMU_HARDDISK_f08311eb-e2ad-4eb7-8da9-ac6d899a2409'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--5368e13a--d21f--5912--91bd--da537815d9e8-osd--block--5368e13a--d21f--5912--91bd--da537815d9e8'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-m0F2nF-j9UB-mZfT-bRv8-xW1c-EVch-ZfDvHt', 'scsi-0QEMU_QEMU_HARDDISK_e03b171b-b829-41a0-a0b6-cc46d85f013d', 'scsi-SQEMU_QEMU_HARDDISK_e03b171b-b829-41a0-a0b6-cc46d85f013d'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_9fa3c153-0e2a-45ef-8e48-e0095eb937ad', 'scsi-SQEMU_QEMU_HARDDISK_9fa3c153-0e2a-45ef-8e48-e0095eb937ad'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-21-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part1', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part14', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part15', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part16', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part1', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part14', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part15', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part16', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--ed04aa0b--4f5a--523b--a5d2--05072a1b260d-osd--block--ed04aa0b--4f5a--523b--a5d2--05072a1b260d'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-nra7ru-O0w6-dM26-lz5o-2arZ-rXsE-lANagX', 'scsi-0QEMU_QEMU_HARDDISK_4e74a544-8b3a-4a05-98a1-b7cc13f13c18', 'scsi-SQEMU_QEMU_HARDDISK_4e74a544-8b3a-4a05-98a1-b7cc13f13c18'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--ede169e1--d974--55ed--a938--c303cb89cb5b-osd--block--ede169e1--d974--55ed--a938--c303cb89cb5b'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-HVhEyD-n8kM-PvHm-CUDD-tuRZ-igjC-Nzour2', 'scsi-0QEMU_QEMU_HARDDISK_f76ae8f0-65cc-4a62-b19c-23f8a40fbde6', 'scsi-SQEMU_QEMU_HARDDISK_f76ae8f0-65cc-4a62-b19c-23f8a40fbde6'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--45b09425--beda--50b4--a83b--1fc50b5adca4-osd--block--45b09425--beda--50b4--a83b--1fc50b5adca4'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-XQTe4H-QTjK-DLSs-Qjj6-RKjC-da4T-xXChra', 'scsi-0QEMU_QEMU_HARDDISK_666e73fe-87f5-4f65-9ff2-95a038c8b18d', 'scsi-SQEMU_QEMU_HARDDISK_666e73fe-87f5-4f65-9ff2-95a038c8b18d'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ae4aabb-0174-4999-a1d0-4bf78cbb2965', 'scsi-SQEMU_QEMU_HARDDISK_8ae4aabb-0174-4999-a1d0-4bf78cbb2965'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a-osd--block--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-YifJdB-JYWE-HhRV-mF1g-0k83-hzWG-ctJZWy', 'scsi-0QEMU_QEMU_HARDDISK_d7bb81ba-6491-4b5a-bb82-f0b49314eac0', 'scsi-SQEMU_QEMU_HARDDISK_d7bb81ba-6491-4b5a-bb82-f0b49314eac0'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-14-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_cdffe5f1-7f0d-460e-a1db-d6846d226144', 'scsi-SQEMU_QEMU_HARDDISK_cdffe5f1-7f0d-460e-a1db-d6846d226144'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-28-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nSaturday 21 March 2026  00:56:05 +0000 (0:00:00.527)       0:00:16.809 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nSaturday 21 March 2026  00:56:06 +0000 (0:00:00.560)       0:00:17.370 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nSaturday 21 March 2026  00:56:06 +0000 (0:00:00.413)       0:00:17.784 ******** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nSaturday 21 March 2026  00:56:07 +0000 (0:00:01.430)       0:00:19.214 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nSaturday 21 March 2026  00:56:08 +0000 (0:00:00.238)       0:00:19.453 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nSaturday 21 March 2026  00:56:08 +0000 (0:00:00.390)       0:00:19.844 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nSaturday 21 March 2026  00:56:08 +0000 (0:00:00.392)       0:00:20.236 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nSaturday 21 March 2026  00:56:09 +0000 (0:00:00.967)       0:00:21.203 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nSaturday 21 March 2026  00:56:10 +0000 (0:00:00.301)       0:00:21.504 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nSaturday 21 March 2026  00:56:10 +0000 (0:00:00.584)       0:00:22.088 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nSaturday 21 March 2026  00:56:11 +0000 (0:00:00.260)       0:00:22.349 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nSaturday 21 March 2026  00:56:11 +0000 (0:00:00.273)       0:00:22.622 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nSaturday 21 March 2026  00:56:11 +0000 (0:00:00.256)       0:00:22.879 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nSaturday 21 March 2026  00:56:12 +0000 (0:00:00.464)       0:00:23.344 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nSaturday 21 March 2026  00:56:12 +0000 (0:00:00.343)       0:00:23.687 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nSaturday 21 March 2026  00:56:12 +0000 (0:00:00.312)       0:00:24.000 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nSaturday 21 March 2026  00:56:13 +0000 (0:00:00.314)       0:00:24.314 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nSaturday 21 March 2026  00:56:13 +0000 (0:00:00.278)       0:00:24.593 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nSaturday 21 March 2026  00:56:13 +0000 (0:00:00.531)       0:00:25.125 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nSaturday 21 March 2026  00:56:14 +0000 (0:00:01.061)       0:00:26.186 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [Include tasks from the ceph-osd role] ************************************\nSaturday 21 March 2026  00:56:17 +0000 (0:00:02.256)       0:00:28.443 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/openstack_config.yml for testbed-node-5\u001b[0m\n\nTASK [create openstack pool(s)] ************************************************\nSaturday 21 March 2026  00:56:17 +0000 (0:00:00.485)       0:00:28.929 ******** \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'backups', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'volumes', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'images', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'metrics', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'vms', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\nTASK [generate keys] ***********************************************************\nSaturday 21 March 2026  00:57:02 +0000 (0:00:45.248)       0:01:14.178 ******** \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> {{ groups[mon_group_name][0] }}]\u001b[0m\n\nTASK [get keys from monitors] **************************************************\nSaturday 21 March 2026  00:57:24 +0000 (0:00:21.264)       0:01:35.442 ******** \n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [copy ceph key(s) if needed] **********************************************\nSaturday 21 March 2026  00:57:36 +0000 (0:00:12.491)       0:01:47.934 ******** \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> {{ item.1 }}]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=25  \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=28  \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=21  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:57:54 +0000 (0:00:17.558)       0:02:05.492 ******** \n=============================================================================== \ncreate openstack pool(s) ----------------------------------------------- 45.25s\ngenerate keys ---------------------------------------------------------- 21.26s\ncopy ceph key(s) if needed --------------------------------------------- 17.56s\nget keys from monitors ------------------------------------------------- 12.49s\nceph-facts : Find a running mon container ------------------------------- 2.79s\nceph-facts : Set_fact ceph_admin_command -------------------------------- 2.26s\nceph-facts : Get current fsid if cluster is already running ------------- 1.60s\nceph-facts : Read osd pool default crush rule --------------------------- 1.43s\nceph-facts : Set_fact ceph_run_cmd -------------------------------------- 1.06s\nceph-facts : Set_fact _monitor_addresses - ipv4 ------------------------- 0.97s\nceph-facts : Check if it is atomic host --------------------------------- 0.84s\nceph-facts : Check if the ceph mon socket is in-use --------------------- 0.84s\nceph-facts : Check if podman binary is present -------------------------- 0.68s\nceph-facts : Set_fact monitor_name ansible_facts['hostname'] ------------ 0.64s\nceph-facts : Import_tasks set_radosgw_address.yml ----------------------- 0.58s\nceph-facts : Check if the ceph conf exists ------------------------------ 0.56s\nceph-facts : Set_fact _container_exec_cmd ------------------------------- 0.55s\nceph-facts : Collect existed devices ------------------------------------ 0.55s\nceph-facts : Set_fact rgw_instances ------------------------------------- 0.53s\nceph-facts : Set_fact devices generate device list when osd_auto_discovery --- 0.53s\n\u001b[32m2026-03-21 00:57:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:57:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:57:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:58:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:58:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:58:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:58:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:58:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:58:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:58:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:58:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:58:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:58:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:58:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:58:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-21 00:58:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:58:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:58:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state SUCCESS\u001b[0m\n\n\nPLAY [Copy ceph keys to the configuration repository] **************************\n\nTASK [Check if ceph keys exist] ************************************************\nSaturday 21 March 2026  00:57:57 +0000 (0:00:00.237)       0:00:00.237 ******** \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.admin.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.nova.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.glance.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.gnocchi.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.manila.keyring)\u001b[0m\n\nTASK [Fetch all ceph keys] *****************************************************\nSaturday 21 March 2026  00:58:02 +0000 (0:00:04.863)       0:00:05.101 ******** \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.admin.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.nova.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.glance.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.gnocchi.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.manila.keyring)\u001b[0m\n\nTASK [Create share directory] **************************************************\nSaturday 21 March 2026  00:58:06 +0000 (0:00:04.223)       0:00:09.324 ******** \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Write ceph keys to the share directory] **********************************\nSaturday 21 March 2026  00:58:08 +0000 (0:00:01.055)       0:00:10.379 ******** \n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.admin.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.nova.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.glance.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.gnocchi.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.manila.keyring)\u001b[0m\n\nTASK [Check if target directories exist] ***************************************\nSaturday 21 March 2026  00:58:21 +0000 (0:00:13.159)       0:00:23.539 ******** \n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/infrastructure/files/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-volume)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-backup)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-backup)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/nova)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/nova)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/glance)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/gnocchi)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/manila)\u001b[0m\n\nTASK [Write ceph keys to the configuration directory] **************************\nSaturday 21 March 2026  00:58:24 +0000 (0:00:03.333)       0:00:26.872 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.admin.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.nova.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.glance.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.gnocchi.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.manila.keyring)\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=6   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:58:31 +0000 (0:00:06.889)       0:00:33.761 ******** \n=============================================================================== \nWrite ceph keys to the share directory --------------------------------- 13.16s\nWrite ceph keys to the configuration directory -------------------------- 6.89s\nCheck if ceph keys exist ------------------------------------------------ 4.86s\nFetch all ceph keys ----------------------------------------------------- 4.22s\nCheck if target directories exist --------------------------------------- 3.33s\nCreate share directory -------------------------------------------------- 1.06s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  00:56:03 +0000 (0:00:00.279)       0:00:00.279 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  00:56:04 +0000 (0:00:00.259)       0:00:00.539 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_keystone_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_keystone_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_keystone_True)\u001b[0m\n\nPLAY [Apply role keystone] *****************************************************\n\nTASK [keystone : include_tasks] ************************************************\nSaturday 21 March 2026  00:56:04 +0000 (0:00:00.274)       0:00:00.813 ******** \n\u001b[0;36mincluded: /ansible/roles/keystone/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [keystone : Ensuring config directories exist] ****************************\nSaturday 21 March 2026  00:56:04 +0000 (0:00:00.608)       0:00:01.422 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [keystone : Check if policies shall be overwritten] ***********************\nSaturday 21 March 2026  00:56:07 +0000 (0:00:02.250)       0:00:03.673 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Set keystone policy file] *************************************\nSaturday 21 March 2026  00:56:07 +0000 (0:00:00.118)       0:00:03.792 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Check if Keystone domain-specific config is supplied] *********\nSaturday 21 March 2026  00:56:07 +0000 (0:00:00.247)       0:00:04.039 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nSaturday 21 March 2026  00:56:08 +0000 (0:00:00.831)       0:00:04.871 ******** \n\u001b[0;36mincluded: /ansible/roles/keystone/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : keystone | Copying over extra CA certificates] *******\nSaturday 21 March 2026  00:56:09 +0000 (0:00:00.583)       0:00:05.455 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : keystone | Copying over backend internal TLS certificate] ***\nSaturday 21 March 2026  00:56:11 +0000 (0:00:02.937)       0:00:08.392 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : keystone | Copying over backend internal TLS key] ****\nSaturday 21 March 2026  00:56:12 +0000 (0:00:00.516)       0:00:08.908 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying over config.json files for services] ******************\nSaturday 21 March 2026  00:56:13 +0000 (0:00:00.893)       0:00:09.802 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [keystone : Copying over keystone.conf] ***********************************\nSaturday 21 March 2026  00:56:17 +0000 (0:00:03.814)       0:00:13.617 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [keystone : Copying keystone-startup script for keystone] *****************\nSaturday 21 March 2026  00:56:22 +0000 (0:00:05.359)       0:00:18.977 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Create Keystone domain-specific config directory] *************\nSaturday 21 March 2026  00:56:24 +0000 (0:00:01.605)       0:00:20.582 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Get file list in custom domains folder] ***********************\nSaturday 21 March 2026  00:56:25 +0000 (0:00:00.959)       0:00:21.542 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying Keystone Domain specific settings] ********************\nSaturday 21 March 2026  00:56:25 +0000 (0:00:00.300)       0:00:21.843 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying over existing policy file] ****************************\nSaturday 21 March 2026  00:56:25 +0000 (0:00:00.299)       0:00:22.143 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nSaturday 21 March 2026  00:56:26 +0000 (0:00:00.561)       0:00:22.705 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying over wsgi-keystone.conf] ******************************\nSaturday 21 March 2026  00:56:26 +0000 (0:00:00.446)       0:00:23.152 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m\n\nTASK [keystone : Checking whether keystone-paste.ini file exists] **************\nSaturday 21 March 2026  00:56:28 +0000 (0:00:01.587)       0:00:24.740 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [keystone : Copying over keystone-paste.ini] ******************************\nSaturday 21 March 2026  00:56:29 +0000 (0:00:00.971)       0:00:25.711 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Generate the required cron jobs for the node] *****************\nSaturday 21 March 2026  00:56:29 +0000 (0:00:00.488)       0:00:26.199 ******** \n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\nTASK [keystone : Set fact with the generated cron jobs for building the crontab later] ***\nSaturday 21 March 2026  00:56:30 +0000 (0:00:01.120)       0:00:27.320 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying files for keystone-fernet] ****************************\nSaturday 21 March 2026  00:56:31 +0000 (0:00:00.435)       0:00:27.756 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\nTASK [keystone : Copying files for keystone-ssh] *******************************\nSaturday 21 March 2026  00:56:39 +0000 (0:00:08.262)       0:00:36.018 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\nTASK [keystone : Check keystone containers] ************************************\nSaturday 21 March 2026  00:56:42 +0000 (0:00:02.496)       0:00:38.514 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nSaturday 21 March 2026  00:56:44 +0000 (0:00:02.072)       0:00:40.586 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Creating keystone database] ***********************************\nSaturday 21 March 2026  00:56:44 +0000 (0:00:00.277)       0:00:40.864 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Creating Keystone database user and setting permissions] ******\nSaturday 21 March 2026  00:56:47 +0000 (0:00:02.939)       0:00:43.803 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Checking for any running keystone_fernet containers] **********\nSaturday 21 March 2026  00:56:49 +0000 (0:00:02.488)       0:00:46.292 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Group nodes where keystone_fernet is running] *****************\nSaturday 21 March 2026  00:56:50 +0000 (0:00:00.787)       0:00:47.080 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Fail if any hosts need bootstrapping and not all hosts targeted] ***\nSaturday 21 March 2026  00:56:50 +0000 (0:00:00.315)       0:00:47.395 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Running Keystone bootstrap container] *************************\nSaturday 21 March 2026  00:56:51 +0000 (0:00:00.312)       0:00:47.708 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Running Keystone fernet bootstrap container] ******************\nSaturday 21 March 2026  00:57:05 +0000 (0:00:14.282)       0:01:01.990 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Flush handlers] ***********************************************\nSaturday 21 March 2026  00:57:16 +0000 (0:00:10.938)       0:01:12.929 ******** \n\nTASK [keystone : Flush handlers] ***********************************************\nSaturday 21 March 2026  00:57:16 +0000 (0:00:00.062)       0:01:12.992 ******** \n\nTASK [keystone : Flush handlers] ***********************************************\nSaturday 21 March 2026  00:57:16 +0000 (0:00:00.064)       0:01:13.056 ******** \n\nRUNNING HANDLER [keystone : Restart keystone-ssh container] ********************\nSaturday 21 March 2026  00:57:16 +0000 (0:00:00.067)       0:01:13.123 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [keystone : Restart keystone-fernet container] *****************\nSaturday 21 March 2026  00:57:26 +0000 (0:00:10.084)       0:01:23.208 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [keystone : Restart keystone container] ************************\nSaturday 21 March 2026  00:57:36 +0000 (0:00:09.863)       0:01:33.071 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nSaturday 21 March 2026  00:57:42 +0000 (0:00:05.760)       0:01:38.832 ******** \n\u001b[0;36mincluded: /ansible/roles/keystone/tasks/distribute_fernet.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [keystone : Waiting for Keystone SSH port to be UP] ***********************\nSaturday 21 March 2026  00:57:43 +0000 (0:00:00.677)       0:01:39.510 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [keystone : Run key distribution] *****************************************\nSaturday 21 March 2026  00:57:43 +0000 (0:00:00.746)       0:01:40.257 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Creating admin project, user, role, service, and endpoint] ****\nSaturday 21 March 2026  00:57:45 +0000 (0:00:01.740)       0:01:41.998 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=RegionOne)\u001b[0m\n\nTASK [service-ks-register : keystone | Creating services] **********************\nSaturday 21 March 2026  00:57:58 +0000 (0:00:12.512)       0:01:54.510 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=keystone (identity))\u001b[0m\n\nTASK [service-ks-register : keystone | Creating endpoints] *********************\nSaturday 21 March 2026  00:58:27 +0000 (0:00:29.143)       0:02:23.653 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=keystone -> https://api-int.testbed.osism.xyz:5000 -> internal)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=keystone -> https://api.testbed.osism.xyz:5000 -> public)\u001b[0m\n\nTASK [service-ks-register : keystone | Creating projects] **********************\nSaturday 21 March 2026  00:58:34 +0000 (0:00:07.179)       0:02:30.833 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [service-ks-register : keystone | Creating users] *************************\nSaturday 21 March 2026  00:58:34 +0000 (0:00:00.119)       0:02:30.953 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [service-ks-register : keystone | Creating roles] *************************\nSaturday 21 March 2026  00:58:34 +0000 (0:00:00.114)       0:02:31.068 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [service-ks-register : keystone | Granting user roles] ********************\nSaturday 21 March 2026  00:58:34 +0000 (0:00:00.119)       0:02:31.187 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Creating default user role] ***********************************\nSaturday 21 March 2026  00:58:35 +0000 (0:00:00.313)       0:02:31.501 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nSaturday 21 March 2026  00:58:38 +0000 (0:00:03.228)       0:02:34.730 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=33  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:58:38 +0000 (0:00:00.551)       0:02:35.281 ******** \n=============================================================================== \nservice-ks-register : keystone | Creating services --------------------- 29.14s\nkeystone : Running Keystone bootstrap container ------------------------ 14.28s\nkeystone : Creating admin project, user, role, service, and endpoint --- 12.51s\nkeystone : Running Keystone fernet bootstrap container ----------------- 10.94s\nkeystone : Restart keystone-ssh container ------------------------------ 10.08s\nkeystone : Restart keystone-fernet container ---------------------------- 9.86s\nkeystone : Copying files for keystone-fernet ---------------------------- 8.26s\nservice-ks-register : keystone | Creating endpoints --------------------- 7.18s\nkeystone : Restart keystone container ----------------------------------- 5.76s\nkeystone : Copying over keystone.conf ----------------------------------- 5.36s\nkeystone : Copying over config.json files for services ------------------ 3.81s\nkeystone : Creating default user role ----------------------------------- 3.23s\nkeystone : Creating keystone database ----------------------------------- 2.94s\nservice-cert-copy : keystone | Copying over extra CA certificates ------- 2.94s\nkeystone : Copying files for keystone-ssh ------------------------------- 2.50s\nkeystone : Creating Keystone database user and setting permissions ------ 2.49s\nkeystone : Ensuring config directories exist ---------------------------- 2.25s\nkeystone : Check keystone containers ------------------------------------ 2.07s\nkeystone : Run key distribution ----------------------------------------- 1.74s\nkeystone : Copying keystone-startup script for keystone ----------------- 1.61s\n\u001b[32m2026-03-21 00:58:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:58:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:58:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:58:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:58:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:58:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:58:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:58:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:59:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:59:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:59:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:59:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:59:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:59:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:59:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:59:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:59:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-21 00:59:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:59:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:59:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:59:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:59:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:59:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:59:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:59:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:59:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:59:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:59:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 00:59:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 00:59:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:00:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:00:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:00:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:00:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:00:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state SUCCESS\u001b[0m\n\n\nPLAY [Apply role cephclient] ***************************************************\n\nTASK [osism.services.cephclient : Include container tasks] *********************\nSaturday 21 March 2026  00:58:35 +0000 (0:00:00.306)       0:00:00.306 ******** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/cephclient/tasks/container.yml for testbed-manager\u001b[0m\n\nTASK [osism.services.cephclient : Create required directories] *****************\nSaturday 21 March 2026  00:58:35 +0000 (0:00:00.241)       0:00:00.547 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/cephclient/configuration)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/cephclient/data)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/cephclient)\u001b[0m\n\nTASK [osism.services.cephclient : Copy configuration files] ********************\nSaturday 21 March 2026  00:58:36 +0000 (0:00:01.557)       0:00:02.105 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item={'src': 'ceph.conf.j2', 'dest': '/opt/cephclient/configuration/ceph.conf'})\u001b[0m\n\nTASK [osism.services.cephclient : Copy keyring file] ***************************\nSaturday 21 March 2026  00:58:38 +0000 (0:00:01.190)       0:00:03.295 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.cephclient : Copy docker-compose.yml file] ****************\nSaturday 21 March 2026  00:58:38 +0000 (0:00:00.867)       0:00:04.163 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.cephclient : Manage cephclient service] *******************\nSaturday 21 March 2026  00:58:39 +0000 (0:00:00.898)       0:00:05.061 ******** \n\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage cephclient service (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.cephclient : Copy wrapper scripts] ************************\nSaturday 21 March 2026  00:59:15 +0000 (0:00:35.560)       0:00:40.622 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph-authtool)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=rados)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=radosgw-admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=rbd)\u001b[0m\n\nTASK [osism.services.cephclient : Remove old wrapper scripts] ******************\nSaturday 21 March 2026  00:59:18 +0000 (0:00:03.550)       0:00:44.173 ******** \n\u001b[0;32mok: [testbed-manager] => (item=crushtool)\u001b[0m\n\nTASK [osism.services.cephclient : Include package tasks] ***********************\nSaturday 21 March 2026  00:59:19 +0000 (0:00:00.514)       0:00:44.687 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.cephclient : Include rook task] ***************************\nSaturday 21 March 2026  00:59:19 +0000 (0:00:00.108)       0:00:44.795 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.cephclient : Restart cephclient service] *******\nSaturday 21 March 2026  00:59:19 +0000 (0:00:00.269)       0:00:45.065 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.cephclient : Ensure that all containers are up] ***\nSaturday 21 March 2026  00:59:21 +0000 (0:00:01.333)       0:00:46.398 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.cephclient : Wait for an healthy service] ******\nSaturday 21 March 2026  00:59:21 +0000 (0:00:00.809)       0:00:47.207 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.cephclient : Copy bash completion scripts] *****\nSaturday 21 March 2026  00:59:22 +0000 (0:00:00.527)       0:00:47.735 ******** \n\u001b[0;32mok: [testbed-manager] => (item=ceph)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=rados)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=radosgw-admin)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=rbd)\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  00:59:23 +0000 (0:00:01.142)       0:00:48.877 ******** \n=============================================================================== \nosism.services.cephclient : Manage cephclient service ------------------ 35.56s\nosism.services.cephclient : Copy wrapper scripts ------------------------ 3.55s\nosism.services.cephclient : Create required directories ----------------- 1.56s\nosism.services.cephclient : Restart cephclient service ------------------ 1.33s\nosism.services.cephclient : Copy configuration files -------------------- 1.19s\nosism.services.cephclient : Copy bash completion scripts ---------------- 1.14s\nosism.services.cephclient : Copy docker-compose.yml file ---------------- 0.90s\nosism.services.cephclient : Copy keyring file --------------------------- 0.87s\nosism.services.cephclient : Ensure that all containers are up ----------- 0.81s\nosism.services.cephclient : Wait for an healthy service ----------------- 0.53s\nosism.services.cephclient : Remove old wrapper scripts ------------------ 0.51s\nosism.services.cephclient : Include rook task --------------------------- 0.27s\nosism.services.cephclient : Include container tasks --------------------- 0.24s\nosism.services.cephclient : Include package tasks ----------------------- 0.11s\n\n\nPLAY [Download ironic ipa images] **********************************************\n\nTASK [Ensure the destination directory exists] *********************************\nSaturday 21 March 2026  00:58:43 +0000 (0:00:00.175)       0:00:00.175 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Download ironic-agent initramfs] *****************************************\nSaturday 21 March 2026  00:58:45 +0000 (0:00:01.423)       0:00:01.598 ******** \n\u001b[1;30mFAILED - RETRYING: [localhost]: Download ironic-agent initramfs (3 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [localhost]: Download ironic-agent initramfs (2 retries left).\u001b[0m\n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Download ironic-agent kernel] ********************************************\nSaturday 21 March 2026  01:00:05 +0000 (0:01:20.183)       0:01:21.782 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  01:00:11 +0000 (0:00:06.172)       0:01:27.955 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  01:00:11 +0000 (0:00:00.302)       0:01:28.257 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_ironic_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_ironic_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_ironic_False)\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: enable_ironic_True\u001b[0m\n\nPLAY [Apply role ironic] *******************************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  01:00:12 +0000 (0:00:00.376)       0:01:28.634 ******** \n=============================================================================== \nDownload ironic-agent initramfs ---------------------------------------- 80.18s\nDownload ironic-agent kernel -------------------------------------------- 6.17s\nEnsure the destination directory exists --------------------------------- 1.42s\nGroup hosts based on enabled services ----------------------------------- 0.38s\nGroup hosts based on Kolla action --------------------------------------- 0.30s\n\u001b[32m2026-03-21 01:00:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:00:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:00:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:00:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:00:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:00:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:00:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:00:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:00:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:00:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:00:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:00:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:00:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:00:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  00:58:43 +0000 (0:00:00.458)       0:00:00.458 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  00:58:43 +0000 (0:00:00.393)       0:00:00.851 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_barbican_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_barbican_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_barbican_True)\u001b[0m\n\nPLAY [Apply role barbican] *****************************************************\n\nTASK [barbican : include_tasks] ************************************************\nSaturday 21 March 2026  00:58:44 +0000 (0:00:00.375)       0:00:01.227 ******** \n\u001b[0;36mincluded: /ansible/roles/barbican/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : barbican | Creating services] **********************\nSaturday 21 March 2026  00:58:45 +0000 (0:00:00.750)       0:00:01.977 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican (key-manager))\u001b[0m\n\nTASK [service-ks-register : barbican | Creating endpoints] *********************\nSaturday 21 March 2026  00:58:49 +0000 (0:00:04.135)       0:00:06.113 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> https://api-int.testbed.osism.xyz:9311 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> https://api.testbed.osism.xyz:9311 -> public)\u001b[0m\n\nTASK [service-ks-register : barbican | Creating projects] **********************\nSaturday 21 March 2026  00:58:56 +0000 (0:00:07.062)       0:00:13.176 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : barbican | Creating users] *************************\nSaturday 21 March 2026  00:58:59 +0000 (0:00:03.627)       0:00:16.803 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : barbican | Creating roles] *************************\nSaturday 21 March 2026  00:59:03 +0000 (0:00:03.738)       0:00:20.541 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=key-manager:service-admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=creator)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=observer)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=audit)\u001b[0m\n\nTASK [service-ks-register : barbican | Granting user roles] ********************\nSaturday 21 March 2026  00:59:20 +0000 (0:00:17.225)       0:00:37.767 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> service -> admin)\u001b[0m\n\nTASK [barbican : Ensuring config directories exist] ****************************\nSaturday 21 March 2026  00:59:25 +0000 (0:00:04.563)       0:00:42.331 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [barbican : Ensuring vassals config directories exist] ********************\nSaturday 21 March 2026  00:59:28 +0000 (0:00:02.996)       0:00:45.327 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican-api/vassals)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=barbican-api/vassals)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=barbican-api/vassals)\u001b[0m\n\nTASK [barbican : Check if policies shall be overwritten] ***********************\nSaturday 21 March 2026  00:59:29 +0000 (0:00:01.443)       0:00:46.770 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [barbican : Set barbican policy file] *************************************\nSaturday 21 March 2026  00:59:30 +0000 (0:00:00.157)       0:00:46.928 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [barbican : include_tasks] ************************************************\nSaturday 21 March 2026  00:59:30 +0000 (0:00:00.236)       0:00:47.165 ******** \n\u001b[0;36mincluded: /ansible/roles/barbican/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : barbican | Copying over extra CA certificates] *******\nSaturday 21 March 2026  00:59:30 +0000 (0:00:00.527)       0:00:47.693 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : barbican | Copying over backend internal TLS certificate] ***\nSaturday 21 March 2026  00:59:35 +0000 (0:00:04.292)       0:00:51.985 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : barbican | Copying over backend internal TLS key] ****\nSaturday 21 March 2026  00:59:36 +0000 (0:00:01.296)       0:00:53.282 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [barbican : Copying over config.json files for services] ******************\nSaturday 21 March 2026  00:59:37 +0000 (0:00:00.759)       0:00:54.041 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [barbican : Copying over barbican-api.ini] ********************************\nSaturday 21 March 2026  00:59:42 +0000 (0:00:05.318)       0:00:59.360 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [barbican : Checking whether barbican-api-paste.ini file exists] **********\nSaturday 21 March 2026  00:59:44 +0000 (0:00:01.844)       0:01:01.204 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [barbican : Copying over barbican-api-paste.ini] **************************\nSaturday 21 March 2026  00:59:46 +0000 (0:00:02.145)       0:01:03.350 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [barbican : Copying over barbican.conf] ***********************************\nSaturday 21 March 2026  00:59:47 +0000 (0:00:00.661)       0:01:04.012 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [barbican : Copying over existing policy file] ****************************\nSaturday 21 March 2026  00:59:58 +0000 (0:00:11.290)       0:01:15.302 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [barbican : Check barbican containers] ************************************\nSaturday 21 March 2026  01:00:00 +0000 (0:00:01.932)       0:01:17.236 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [barbican : include_tasks] ************************************************\nSaturday 21 March 2026  01:00:03 +0000 (0:00:03.462)       0:01:20.699 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [barbican : Creating barbican database] ***********************************\nSaturday 21 March 2026  01:00:04 +0000 (0:00:00.813)       0:01:21.512 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [barbican : Creating barbican database user and setting permissions] ******\nSaturday 21 March 2026  01:00:07 +0000 (0:00:02.377)       0:01:23.889 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [barbican : Running barbican bootstrap container] *************************\nSaturday 21 March 2026  01:00:09 +0000 (0:00:02.365)       0:01:26.254 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [barbican : Flush handlers] ***********************************************\nSaturday 21 March 2026  01:00:20 +0000 (0:00:11.433)       0:01:37.688 ******** \n\nTASK [barbican : Flush handlers] ***********************************************\nSaturday 21 March 2026  01:00:21 +0000 (0:00:00.339)       0:01:38.028 ******** \n\nTASK [barbican : Flush handlers] ***********************************************\nSaturday 21 March 2026  01:00:21 +0000 (0:00:00.068)       0:01:38.096 ******** \n\nRUNNING HANDLER [barbican : Restart barbican-api container] ********************\nSaturday 21 March 2026  01:00:21 +0000 (0:00:00.065)       0:01:38.162 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [barbican : Restart barbican-keystone-listener container] ******\nSaturday 21 March 2026  01:00:31 +0000 (0:00:10.300)       0:01:48.462 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [barbican : Restart barbican-worker container] *****************\nSaturday 21 March 2026  01:00:41 +0000 (0:00:10.064)       0:01:58.527 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  01:00:52 +0000 (0:00:10.839)       0:02:09.367 ******** \n=============================================================================== \nservice-ks-register : barbican | Creating roles ------------------------ 17.23s\nbarbican : Running barbican bootstrap container ------------------------ 11.43s\nbarbican : Copying over barbican.conf ---------------------------------- 11.29s\nbarbican : Restart barbican-worker container --------------------------- 10.84s\nbarbican : Restart barbican-api container ------------------------------ 10.30s\nbarbican : Restart barbican-keystone-listener container ---------------- 10.06s\nservice-ks-register : barbican | Creating endpoints --------------------- 7.06s\nbarbican : Copying over config.json files for services ------------------ 5.32s\nservice-ks-register : barbican | Granting user roles -------------------- 4.56s\nservice-cert-copy : barbican | Copying over extra CA certificates ------- 4.29s\nservice-ks-register : barbican | Creating services ---------------------- 4.14s\nservice-ks-register : barbican | Creating users ------------------------- 3.74s\nservice-ks-register : barbican | Creating projects ---------------------- 3.63s\nbarbican : Check barbican containers ------------------------------------ 3.46s\nbarbican : Ensuring config directories exist ---------------------------- 3.00s\nbarbican : Creating barbican database ----------------------------------- 2.38s\nbarbican : Creating barbican database user and setting permissions ------ 2.37s\nbarbican : Checking whether barbican-api-paste.ini file exists ---------- 2.15s\nbarbican : Copying over existing policy file ---------------------------- 1.93s\nbarbican : Copying over barbican-api.ini -------------------------------- 1.84s\n\u001b[32m2026-03-21 01:00:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:00:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:00:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:01:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-21 01:01:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:01:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:01:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:01:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:01:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:01:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state SUCCESS\u001b[0m\n\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Bootstraph ceph dashboard] ***********************************************\n\nTASK [Disable the ceph dashboard] **********************************************\nSaturday 21 March 2026  00:59:27 +0000 (0:00:00.191)       0:00:00.191 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/ssl to false] ******************************************\nSaturday 21 March 2026  00:59:30 +0000 (0:00:02.428)       0:00:02.619 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/server_port to 7000] ***********************************\nSaturday 21 March 2026  00:59:31 +0000 (0:00:00.908)       0:00:03.527 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/server_addr to 0.0.0.0] ********************************\nSaturday 21 March 2026  00:59:31 +0000 (0:00:00.842)       0:00:04.370 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/standby_behaviour to error] ****************************\nSaturday 21 March 2026  00:59:32 +0000 (0:00:01.064)       0:00:05.435 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/standby_error_status_code to 404] **********************\nSaturday 21 March 2026  00:59:34 +0000 (0:00:01.278)       0:00:06.713 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Enable the ceph dashboard] ***********************************************\nSaturday 21 March 2026  00:59:35 +0000 (0:00:01.151)       0:00:07.865 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Write ceph_dashboard_password to temporary file] *************************\nSaturday 21 March 2026  00:59:37 +0000 (0:00:02.012)       0:00:09.877 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Create admin user] *******************************************************\nSaturday 21 March 2026  00:59:38 +0000 (0:00:01.156)       0:00:11.033 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Remove temporary file for ceph_dashboard_password] ***********************\nSaturday 21 March 2026  01:00:35 +0000 (0:00:56.532)       0:01:07.566 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nPLAY [Restart ceph manager services] *******************************************\n\nTASK [Restart ceph manager service] ********************************************\nSaturday 21 March 2026  01:00:35 +0000 (0:00:00.133)       0:01:07.699 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY [Restart ceph manager services] *******************************************\n\nTASK [Restart ceph manager service] ********************************************\nSaturday 21 March 2026  01:00:46 +0000 (0:00:11.530)       0:01:19.229 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY [Restart ceph manager services] *******************************************\n\nTASK [Restart ceph manager service] ********************************************\nSaturday 21 March 2026  01:00:48 +0000 (0:00:01.299)       0:01:20.529 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  01:00:59 +0000 (0:00:11.323)       0:01:31.852 ******** \n=============================================================================== \nCreate admin user ------------------------------------------------------ 56.53s\nRestart ceph manager service ------------------------------------------- 24.15s\nDisable the ceph dashboard ---------------------------------------------- 2.43s\nEnable the ceph dashboard ----------------------------------------------- 2.01s\nSet mgr/dashboard/standby_behaviour to error ---------------------------- 1.28s\nWrite ceph_dashboard_password to temporary file ------------------------- 1.16s\nSet mgr/dashboard/standby_error_status_code to 404 ---------------------- 1.15s\nSet mgr/dashboard/server_addr to 0.0.0.0 -------------------------------- 1.06s\nSet mgr/dashboard/ssl to false ------------------------------------------ 0.91s\nSet mgr/dashboard/server_port to 7000 ----------------------------------- 0.84s\nRemove temporary file for ceph_dashboard_password ----------------------- 0.13s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  01:00:17 +0000 (0:00:00.413)       0:00:00.413 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  01:00:17 +0000 (0:00:00.256)       0:00:00.669 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_placement_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_placement_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_placement_True)\u001b[0m\n\nPLAY [Apply role placement] ****************************************************\n\nTASK [placement : include_tasks] ***********************************************\nSaturday 21 March 2026  01:00:18 +0000 (0:00:00.276)       0:00:00.945 ******** \n\u001b[0;36mincluded: /ansible/roles/placement/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : placement | Creating services] *********************\nSaturday 21 March 2026  01:00:19 +0000 (0:00:01.016)       0:00:01.962 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=placement (placement))\u001b[0m\n\nTASK [service-ks-register : placement | Creating endpoints] ********************\nSaturday 21 March 2026  01:00:22 +0000 (0:00:03.374)       0:00:05.336 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> https://api-int.testbed.osism.xyz:8780 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> https://api.testbed.osism.xyz:8780 -> public)\u001b[0m\n\nTASK [service-ks-register : placement | Creating projects] *********************\nSaturday 21 March 2026  01:00:28 +0000 (0:00:05.813)       0:00:11.150 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : placement | Creating users] ************************\nSaturday 21 March 2026  01:00:32 +0000 (0:00:03.763)       0:00:14.914 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : placement | Creating roles] ************************\nSaturday 21 March 2026  01:00:35 +0000 (0:00:03.720)       0:00:18.634 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : placement | Granting user roles] *******************\nSaturday 21 March 2026  01:00:38 +0000 (0:00:03.049)       0:00:21.684 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> service -> admin)\u001b[0m\n\nTASK [placement : include_tasks] ***********************************************\nSaturday 21 March 2026  01:00:43 +0000 (0:00:04.235)       0:00:25.919 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [placement : Ensuring config directories exist] ***************************\nSaturday 21 March 2026  01:00:43 +0000 (0:00:00.527)       0:00:26.447 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [placement : Check if policies shall be overwritten] **********************\nSaturday 21 March 2026  01:00:45 +0000 (0:00:02.113)       0:00:28.560 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [placement : Set placement policy file] ***********************************\nSaturday 21 March 2026  01:00:45 +0000 (0:00:00.124)       0:00:28.685 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [placement : include_tasks] ***********************************************\nSaturday 21 March 2026  01:00:46 +0000 (0:00:00.254)       0:00:28.939 ******** \n\u001b[0;36mincluded: /ansible/roles/placement/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : placement | Copying over extra CA certificates] ******\nSaturday 21 March 2026  01:00:47 +0000 (0:00:01.085)       0:00:30.025 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [service-cert-copy : placement | Copying over backend internal TLS certificate] ***\nSaturday 21 March 2026  01:00:49 +0000 (0:00:01.991)       0:00:32.016 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : placement | Copying over backend internal TLS key] ***\nSaturday 21 March 2026  01:00:50 +0000 (0:00:01.008)       0:00:33.024 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [placement : Copying over config.json files for services] *****************\nSaturday 21 March 2026  01:00:51 +0000 (0:00:01.297)       0:00:34.322 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [placement : Copying over placement.conf] *********************************\nSaturday 21 March 2026  01:00:53 +0000 (0:00:01.735)       0:00:36.058 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [placement : Copying over placement-api wsgi configuration] ***************\nSaturday 21 March 2026  01:00:55 +0000 (0:00:02.449)       0:00:38.508 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2)\u001b[0m\n\nTASK [placement : Copying over migrate-db.rc.j2 configuration] *****************\nSaturday 21 March 2026  01:00:57 +0000 (0:00:01.622)       0:00:40.130 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [placement : Copying over existing policy file] ***************************\nSaturday 21 March 2026  01:00:58 +0000 (0:00:01.502)       0:00:41.632 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [placement : Check placement containers] **********************************\nSaturday 21 March 2026  01:01:00 +0000 (0:00:01.687)       0:00:43.319 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [placement : Creating placement databases] ********************************\nSaturday 21 March 2026  01:01:02 +0000 (0:00:01.461)       0:00:44.780 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [placement : Creating placement databases user and setting permissions] ***\nSaturday 21 March 2026  01:01:03 +0000 (0:00:01.960)       0:00:46.741 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [placement : Running placement bootstrap container] ***********************\nSaturday 21 March 2026  01:01:06 +0000 (0:00:02.140)       0:00:48.881 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [placement : Flush handlers] **********************************************\nSaturday 21 March 2026  01:01:18 +0000 (0:00:12.443)       0:01:01.325 ******** \n\nTASK [placement : Flush handlers] **********************************************\nSaturday 21 March 2026  01:01:18 +0000 (0:00:00.056)       0:01:01.381 ******** \n\nTASK [placement : Flush handlers] **********************************************\nSaturday 21 March 2026  01:01:18 +0000 (0:00:00.059)       0:01:01.441 ******** \n\nRUNNING HANDLER [placement : Restart placement-api container] ******************\nSaturday 21 March 2026  01:01:18 +0000 (0:00:00.060)       0:01:01.501 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  01:01:28 +0000 (0:00:09.418)       0:01:10.920 ******** \n=============================================================================== \nplacement : Running placement bootstrap container ---------------------- 12.44s\nplacement : Restart placement-api container ----------------------------- 9.42s\nservice-ks-register : placement | Creating endpoints -------------------- 5.81s\nservice-ks-register : placement | Granting user roles ------------------- 4.24s\nservice-ks-register : placement | Creating projects --------------------- 3.76s\nservice-ks-register : placement | Creating users ------------------------ 3.72s\nservice-ks-register : placement | Creating services --------------------- 3.37s\nservice-ks-register : placement | Creating roles ------------------------ 3.05s\nplacement : Copying over placement.conf --------------------------------- 2.45s\nplacement : Creating placement databases user and setting permissions --- 2.14s\nplacement : Ensuring config directories exist --------------------------- 2.11s\nservice-cert-copy : placement | Copying over extra CA certificates ------ 1.99s\nplacement : Creating placement databases -------------------------------- 1.96s\nplacement : Copying over config.json files for services ----------------- 1.74s\nplacement : Copying over existing policy file --------------------------- 1.69s\nplacement : Copying over placement-api wsgi configuration --------------- 1.62s\nplacement : Copying over migrate-db.rc.j2 configuration ----------------- 1.50s\nplacement : Check placement containers ---------------------------------- 1.46s\nservice-cert-copy : placement | Copying over backend internal TLS key --- 1.29s\nplacement : include_tasks ----------------------------------------------- 1.09s\n\u001b[32m2026-03-21 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 29d2a090-e26a-46c7-8734-92002c46d499 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:01:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 29d2a090-e26a-46c7-8734-92002c46d499 is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-21 01:01:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:01:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:01:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:01:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  01:01:31 +0000 (0:00:00.167)       0:00:00.167 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  01:01:31 +0000 (0:00:00.296)       0:00:00.464 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_keystone_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_keystone_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_keystone_True)\u001b[0m\n\nPLAY [Wait for the Keystone service] *******************************************\n\nTASK [Waiting for Keystone public port to be UP] *******************************\nSaturday 21 March 2026  01:01:31 +0000 (0:00:00.418)       0:00:00.882 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  01:01:32 +0000 (0:00:00.886)       0:00:01.768 ******** \n=============================================================================== \nWaiting for Keystone public port to be UP ------------------------------- 0.89s\nGroup hosts based on enabled services ----------------------------------- 0.42s\nGroup hosts based on Kolla action --------------------------------------- 0.30s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  00:58:43 +0000 (0:00:00.777)       0:00:00.778 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  00:58:44 +0000 (0:00:00.423)       0:00:01.202 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_designate_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_designate_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_designate_True)\u001b[0m\n\nPLAY [Apply role designate] ****************************************************\n\nTASK [designate : include_tasks] ***********************************************\nSaturday 21 March 2026  00:58:44 +0000 (0:00:00.441)       0:00:01.643 ******** \n\u001b[0;36mincluded: /ansible/roles/designate/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : designate | Creating services] *********************\nSaturday 21 March 2026  00:58:45 +0000 (0:00:00.959)       0:00:02.603 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=designate (dns))\u001b[0m\n\nTASK [service-ks-register : designate | Creating endpoints] ********************\nSaturday 21 March 2026  00:58:49 +0000 (0:00:04.012)       0:00:06.616 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> https://api-int.testbed.osism.xyz:9001 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> https://api.testbed.osism.xyz:9001 -> public)\u001b[0m\n\nTASK [service-ks-register : designate | Creating projects] *********************\nSaturday 21 March 2026  00:58:56 +0000 (0:00:07.153)       0:00:13.769 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : designate | Creating users] ************************\nSaturday 21 March 2026  00:59:00 +0000 (0:00:03.304)       0:00:17.073 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : designate | Creating roles] ************************\nSaturday 21 March 2026  00:59:04 +0000 (0:00:04.272)       0:00:21.346 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : designate | Granting user roles] *******************\nSaturday 21 March 2026  00:59:08 +0000 (0:00:03.648)       0:00:24.994 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> service -> admin)\u001b[0m\n\nTASK [designate : Ensuring config directories exist] ***************************\nSaturday 21 March 2026  00:59:11 +0000 (0:00:03.672)       0:00:28.667 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : Check if policies shall be overwritten] **********************\nSaturday 21 March 2026  00:59:14 +0000 (0:00:03.062)       0:00:31.729 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [designate : Set designate policy file] ***********************************\nSaturday 21 March 2026  00:59:15 +0000 (0:00:00.116)       0:00:31.846 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [designate : include_tasks] ***********************************************\nSaturday 21 March 2026  00:59:15 +0000 (0:00:00.243)       0:00:32.089 ******** \n\u001b[0;36mincluded: /ansible/roles/designate/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : designate | Copying over extra CA certificates] ******\nSaturday 21 March 2026  00:59:15 +0000 (0:00:00.454)       0:00:32.544 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : designate | Copying over backend internal TLS certificate] ***\nSaturday 21 March 2026  00:59:22 +0000 (0:00:06.700)       0:00:39.245 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : designate | Copying over backend internal TLS key] ***\nSaturday 21 March 2026  00:59:23 +0000 (0:00:00.890)       0:00:40.135 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [designate : Copying over config.json files for services] *****************\nSaturday 21 March 2026  00:59:24 +0000 (0:00:00.908)       0:00:41.044 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : Copying over designate.conf] *********************************\nSaturday 21 March 2026  00:59:32 +0000 (0:00:08.447)       0:00:49.491 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : Copying over pools.yaml] *************************************\nSaturday 21 March 2026  00:59:55 +0000 (0:00:22.578)       0:01:12.069 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m\n\nTASK [designate : Copying over named.conf] *************************************\nSaturday 21 March 2026  01:00:02 +0000 (0:00:07.699)       0:01:19.769 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m\n\nTASK [designate : Copying over rndc.conf] **************************************\nSaturday 21 March 2026  01:00:06 +0000 (0:00:03.243)       0:01:23.012 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : Copying over rndc.key] ***************************************\nSaturday 21 March 2026  01:00:09 +0000 (0:00:03.611)       0:01:26.624 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : include_tasks] ***********************************************\nSaturday 21 March 2026  01:00:13 +0000 (0:00:03.702)       0:01:30.327 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [designate : Copying over existing policy file] ***************************\nSaturday 21 March 2026  01:00:14 +0000 (0:00:00.494)       0:01:30.821 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [designate : Check designate containers] **********************************\nSaturday 21 March 2026  01:00:15 +0000 (0:00:01.300)       0:01:32.122 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : include_tasks] ***********************************************\nSaturday 21 March 2026  01:00:19 +0000 (0:00:04.514)       0:01:36.637 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [designate : Creating Designate databases] ********************************\nSaturday 21 March 2026  01:00:20 +0000 (0:00:00.383)       0:01:37.020 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=designate)\u001b[0m\n\nTASK [designate : Creating Designate databases user and setting permissions] ***\nSaturday 21 March 2026  01:00:22 +0000 (0:00:01.977)       0:01:38.997 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0 -> {{ groups['designate-central'][0] }}]\u001b[0m\n\nTASK [designate : Running Designate bootstrap container] ***********************\nSaturday 21 March 2026  01:00:24 +0000 (0:00:02.069)       0:01:41.067 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [designate : Flush handlers] **********************************************\nSaturday 21 March 2026  01:00:37 +0000 (0:00:13.328)       0:01:54.395 ******** \n\nTASK [designate : Flush handlers] **********************************************\nSaturday 21 March 2026  01:00:37 +0000 (0:00:00.133)       0:01:54.529 ******** \n\nTASK [designate : Flush handlers] **********************************************\nSaturday 21 March 2026  01:00:37 +0000 (0:00:00.067)       0:01:54.597 ******** \n\nRUNNING HANDLER [designate : Restart designate-backend-bind9 container] ********\nSaturday 21 March 2026  01:00:37 +0000 (0:00:00.063)       0:01:54.661 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-api container] ******************\nSaturday 21 March 2026  01:00:47 +0000 (0:00:10.073)       0:02:04.734 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-central container] **************\nSaturday 21 March 2026  01:00:59 +0000 (0:00:11.329)       0:02:16.063 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-producer container] *************\nSaturday 21 March 2026  01:01:05 +0000 (0:00:05.900)       0:02:21.964 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-mdns container] *****************\nSaturday 21 March 2026  01:01:13 +0000 (0:00:08.278)       0:02:30.243 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-worker container] ***************\nSaturday 21 March 2026  01:01:23 +0000 (0:00:09.924)       0:02:40.167 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [designate : Non-destructive DNS pools update] ****************************\nSaturday 21 March 2026  01:01:33 +0000 (0:00:10.134)       0:02:50.301 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=29  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  01:01:40 +0000 (0:00:06.906)       0:02:57.208 ******** \n=============================================================================== \ndesignate : Copying over designate.conf -------------------------------- 22.58s\ndesignate : Running Designate bootstrap container ---------------------- 13.33s\ndesignate : Restart designate-api container ---------------------------- 11.33s\ndesignate : Restart designate-worker container ------------------------- 10.13s\ndesignate : Restart designate-backend-bind9 container ------------------ 10.08s\ndesignate : Restart designate-mdns container ---------------------------- 9.92s\ndesignate : Copying over config.json files for services ----------------- 8.45s\ndesignate : Restart designate-producer container ------------------------ 8.28s\ndesignate : Copying over pools.yaml ------------------------------------- 7.70s\nservice-ks-register : designate | Creating endpoints -------------------- 7.15s\ndesignate : Non-destructive DNS pools update ---------------------------- 6.91s\nservice-cert-copy : designate | Copying over extra CA certificates ------ 6.70s\ndesignate : Restart designate-central container ------------------------- 5.90s\ndesignate : Check designate containers ---------------------------------- 4.51s\nservice-ks-register : designate | Creating users ------------------------ 4.27s\nservice-ks-register : designate | Creating services --------------------- 4.01s\ndesignate : Copying over rndc.key --------------------------------------- 3.71s\nservice-ks-register : designate | Granting user roles ------------------- 3.67s\nservice-ks-register : designate | Creating roles ------------------------ 3.65s\ndesignate : Copying over rndc.conf -------------------------------------- 3.61s\n\u001b[32m2026-03-21 01:01:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:01:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:01:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:01:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:01:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:01:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:01:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:02:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:02:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:02:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:02:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:02:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-21 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  01:01:43 +0000 (0:00:00.245)       0:00:00.245 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  01:01:44 +0000 (0:00:01.560)       0:00:01.805 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=enable_ceph_rgw_True)\u001b[0m\n\nPLAY [Apply role ceph-rgw] *****************************************************\n\nTASK [ceph-rgw : include_tasks] ************************************************\nSaturday 21 March 2026  01:01:46 +0000 (0:00:01.259)       0:00:03.065 ******** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating services] **********************\nSaturday 21 March 2026  01:01:48 +0000 (0:00:02.084)       0:00:05.150 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=swift (object-store))\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating endpoints] *********************\nSaturday 21 March 2026  01:01:51 +0000 (0:00:03.696)       0:00:08.846 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=swift -> https://api-int.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=swift -> https://api.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s -> public)\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating projects] **********************\nSaturday 21 March 2026  01:01:58 +0000 (0:00:06.164)       0:00:15.011 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating users] *************************\nSaturday 21 March 2026  01:02:01 +0000 (0:00:03.561)       0:00:18.573 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=ceph_rgw -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating roles] *************************\nSaturday 21 March 2026  01:02:05 +0000 (0:00:04.273)       0:00:22.846 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=ResellerAdmin)\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Granting user roles] ********************\nSaturday 21 March 2026  01:02:12 +0000 (0:00:06.495)       0:00:29.341 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=ceph_rgw -> service -> admin)\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  01:02:17 +0000 (0:00:05.043)       0:00:34.384 ******** \n=============================================================================== \nservice-ks-register : ceph-rgw | Creating roles ------------------------- 6.49s\nservice-ks-register : ceph-rgw | Creating endpoints --------------------- 6.16s\nservice-ks-register : ceph-rgw | Granting user roles -------------------- 5.04s\nservice-ks-register : ceph-rgw | Creating users ------------------------- 4.27s\nservice-ks-register : ceph-rgw | Creating services ---------------------- 3.70s\nservice-ks-register : ceph-rgw | Creating projects ---------------------- 3.56s\nceph-rgw : include_tasks ------------------------------------------------ 2.08s\nGroup hosts based on Kolla action --------------------------------------- 1.56s\nGroup hosts based on enabled services ----------------------------------- 1.26s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  01:00:57 +0000 (0:00:00.570)       0:00:00.570 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  01:00:58 +0000 (0:00:00.636)       0:00:01.207 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_magnum_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_magnum_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_magnum_True)\u001b[0m\n\nPLAY [Apply role magnum] *******************************************************\n\nTASK [magnum : include_tasks] **************************************************\nSaturday 21 March 2026  01:00:58 +0000 (0:00:00.377)       0:00:01.584 ******** \n\u001b[0;36mincluded: /ansible/roles/magnum/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : magnum | Creating services] ************************\nSaturday 21 March 2026  01:00:59 +0000 (0:00:00.962)       0:00:02.547 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum (container-infra))\u001b[0m\n\nTASK [service-ks-register : magnum | Creating endpoints] ***********************\nSaturday 21 March 2026  01:01:03 +0000 (0:00:04.335)       0:00:06.882 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> https://api-int.testbed.osism.xyz:9511/v1 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> https://api.testbed.osism.xyz:9511/v1 -> public)\u001b[0m\n\nTASK [service-ks-register : magnum | Creating projects] ************************\nSaturday 21 March 2026  01:01:09 +0000 (0:00:06.159)       0:00:13.042 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : magnum | Creating users] ***************************\nSaturday 21 March 2026  01:01:13 +0000 (0:00:03.527)       0:00:16.570 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : magnum | Creating roles] ***************************\nSaturday 21 March 2026  01:01:16 +0000 (0:00:03.418)       0:00:19.988 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : magnum | Granting user roles] **********************\nSaturday 21 March 2026  01:01:19 +0000 (0:00:02.738)       0:00:22.727 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> service -> admin)\u001b[0m\n\nTASK [magnum : Creating Magnum trustee domain] *********************************\nSaturday 21 March 2026  01:01:23 +0000 (0:00:03.864)       0:00:26.591 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Creating Magnum trustee user] ***********************************\nSaturday 21 March 2026  01:01:26 +0000 (0:00:03.160)       0:00:29.751 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Creating Magnum trustee user role] ******************************\nSaturday 21 March 2026  01:01:30 +0000 (0:00:03.504)       0:00:33.256 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Ensuring config directories exist] ******************************\nSaturday 21 March 2026  01:01:33 +0000 (0:00:03.333)       0:00:36.589 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [magnum : Check if policies shall be overwritten] *************************\nSaturday 21 March 2026  01:01:35 +0000 (0:00:01.674)       0:00:38.264 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Set magnum policy file] *****************************************\nSaturday 21 March 2026  01:01:35 +0000 (0:00:00.134)       0:00:38.398 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [magnum : Check if kubeconfig file is supplied] ***************************\nSaturday 21 March 2026  01:01:35 +0000 (0:00:00.296)       0:00:38.695 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [magnum : Copying over kubeconfig file] ***********************************\nSaturday 21 March 2026  01:01:36 +0000 (0:00:00.845)       0:00:39.540 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [magnum : Set magnum kubeconfig file's path] ******************************\nSaturday 21 March 2026  01:01:38 +0000 (0:00:01.955)       0:00:41.496 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [magnum : include_tasks] **************************************************\nSaturday 21 March 2026  01:01:38 +0000 (0:00:00.342)       0:00:41.838 ******** \n\u001b[0;36mincluded: /ansible/roles/magnum/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : magnum | Copying over extra CA certificates] *********\nSaturday 21 March 2026  01:01:39 +0000 (0:00:00.926)       0:00:42.765 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : magnum | Copying over backend internal TLS certificate] ***\nSaturday 21 March 2026  01:01:42 +0000 (0:00:02.445)       0:00:45.211 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : magnum | Copying over backend internal TLS key] ******\nSaturday 21 March 2026  01:01:43 +0000 (0:00:01.078)       0:00:46.289 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}\u001b[32m2026-03-21 01:02:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-21 01:02:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [magnum : Copying over config.json files for services] ********************\nSaturday 21 March 2026  01:01:44 +0000 (0:00:01.347)       0:00:47.637 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [magnum : Copying over magnum.conf] ***************************************\nSaturday 21 March 2026  01:01:47 +0000 (0:00:03.124)       0:00:50.761 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [magnum : Copying over existing policy file] ******************************\nSaturday 21 March 2026  01:01:55 +0000 (0:00:07.861)       0:00:58.622 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [magnum : Check magnum containers] ****************************************\nSaturday 21 March 2026  01:01:56 +0000 (0:00:01.153)       0:00:59.776 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [magnum : include_tasks] **************************************************\nSaturday 21 March 2026  01:01:58 +0000 (0:00:02.000)       0:01:01.776 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [magnum : Creating Magnum database] ***************************************\nSaturday 21 March 2026  01:01:58 +0000 (0:00:00.306)       0:01:02.083 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Creating Magnum database user and setting permissions] **********\nSaturday 21 March 2026  01:02:01 +0000 (0:00:02.386)       0:01:04.469 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Running Magnum bootstrap container] *****************************\nSaturday 21 March 2026  01:02:04 +0000 (0:00:02.693)       0:01:07.162 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Flush handlers] *************************************************\nSaturday 21 March 2026  01:02:21 +0000 (0:00:17.509)       0:01:24.671 ******** \n\nTASK [magnum : Flush handlers] *************************************************\nSaturday 21 March 2026  01:02:21 +0000 (0:00:00.201)       0:01:24.873 ******** \n\nTASK [magnum : Flush handlers] *************************************************\nSaturday 21 March 2026  01:02:21 +0000 (0:00:00.060)       0:01:24.933 ******** \n\nRUNNING HANDLER [magnum : Restart magnum-api container] ************************\nSaturday 21 March 2026  01:02:21 +0000 (0:00:00.076)       0:01:25.010 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [magnum : Restart magnum-conductor container] ******************\nSaturday 21 March 2026  01:02:37 +0000 (0:00:15.419)       0:01:40.430 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  01:02:45 +0000 (0:00:07.918)       0:01:48.348 ******** \n=============================================================================== \nmagnum : Running Magnum bootstrap container ---------------------------- 17.51s\nmagnum : Restart magnum-api container ---------------------------------- 15.42s\nmagnum : Restart magnum-conductor container ----------------------------- 7.92s\nmagnum : Copying over magnum.conf --------------------------------------- 7.86s\nservice-ks-register : magnum | Creating endpoints ----------------------- 6.16s\nservice-ks-register : magnum | Creating services ------------------------ 4.34s\nservice-ks-register : magnum | Granting user roles ---------------------- 3.86s\nservice-ks-register : magnum | Creating projects ------------------------ 3.53s\nmagnum : Creating Magnum trustee user ----------------------------------- 3.50s\nservice-ks-register : magnum | Creating users --------------------------- 3.42s\nmagnum : Creating Magnum trustee user role ------------------------------ 3.33s\nmagnum : Creating Magnum trustee domain --------------------------------- 3.16s\nmagnum : Copying over config.json files for services -------------------- 3.12s\nservice-ks-register : magnum | Creating roles --------------------------- 2.74s\nmagnum : Creating Magnum database user and setting permissions ---------- 2.69s\nservice-cert-copy : magnum | Copying over extra CA certificates --------- 2.45s\nmagnum : Creating Magnum database --------------------------------------- 2.39s\nmagnum : Check magnum containers ---------------------------------------- 2.00s\nmagnum : Copying over kubeconfig file ----------------------------------- 1.95s\nmagnum : Ensuring config directories exist ------------------------------ 1.67s\n\u001b[32m2026-03-21 01:02:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:02:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:02:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  00:58:44 +0000 (0:00:00.394)       0:00:00.394 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  00:58:45 +0000 (0:00:00.951)       0:00:01.346 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_neutron_True)\u001b[0m\n\nPLAY [Apply role neutron] ******************************************************\n\nTASK [neutron : include_tasks] *************************************************\nSaturday 21 March 2026  00:58:46 +0000 (0:00:00.909)       0:00:02.255 ******** \n\u001b[0;36mincluded: /ansible/roles/neutron/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [neutron : Get container facts] *******************************************\nSaturday 21 March 2026  00:58:47 +0000 (0:00:01.182)       0:00:03.438 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Get container volume facts] ************************************\nSaturday 21 March 2026  00:58:48 +0000 (0:00:01.362)       0:00:04.800 ******** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Check for ML2/OVN presence] ************************************\nSaturday 21 March 2026  00:58:50 +0000 (0:00:01.127)       0:00:05.927 ******** \n\u001b[0;32mok: [testbed-node-0] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [neutron : Check for ML2/OVS presence] ************************************\nSaturday 21 March 2026  00:58:50 +0000 (0:00:00.579)       0:00:06.507 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [service-ks-register : neutron | Creating services] ***********************\nSaturday 21 March 2026  00:58:51 +0000 (0:00:00.613)       0:00:07.121 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron (network))\u001b[0m\n\nTASK [service-ks-register : neutron | Creating endpoints] **********************\nSaturday 21 March 2026  00:58:55 +0000 (0:00:03.942)       0:00:11.063 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> https://api-int.testbed.osism.xyz:9696 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> https://api.testbed.osism.xyz:9696 -> public)\u001b[0m\n\nTASK [service-ks-register : neutron | Creating projects] ***********************\nSaturday 21 March 2026  00:59:01 +0000 (0:00:06.006)       0:00:17.069 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : neutron | Creating users] **************************\nSaturday 21 March 2026  00:59:04 +0000 (0:00:03.309)       0:00:20.379 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : neutron | Creating roles] **************************\nSaturday 21 March 2026  00:59:08 +0000 (0:00:04.289)       0:00:24.669 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : neutron | Granting user roles] *********************\nSaturday 21 March 2026  00:59:11 +0000 (0:00:03.228)       0:00:27.897 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service -> admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service -> service)\u001b[0m\n\nTASK [neutron : include_tasks] *************************************************\nSaturday 21 March 2026  00:59:19 +0000 (0:00:07.300)       0:00:35.198 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Load and persist kernel modules] *****************************************\nSaturday 21 March 2026  00:59:19 +0000 (0:00:00.463)       0:00:35.661 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Check IPv6 support] ********************************************\nSaturday 21 March 2026  00:59:22 +0000 (0:00:02.679)       0:00:38.340 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Setting sysctl values] ***************************************************\nSaturday 21 March 2026  00:59:23 +0000 (0:00:01.464)       0:00:39.805 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Ensuring config directories exist] *****************************\nSaturday 21 March 2026  00:59:27 +0000 (0:00:03.197)       0:00:43.002 ******** \n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\nTASK [neutron : Check if extra ml2 plugins exists] *****************************\nSaturday 21 March 2026  00:59:29 +0000 (0:00:02.811)       0:00:45.814 ******** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/neutron/plugins/' path\u001b[0m\n\u001b[1;35mdue to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/neutron/plugins/' is not\u001b[0m\n\u001b[1;35ma directory\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [neutron : include_tasks] *************************************************\nSaturday 21 March 2026  00:59:30 +0000 (0:00:00.869)       0:00:46.684 ******** \n\u001b[0;36mincluded: /ansible/roles/neutron/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [service-cert-copy : neutron | Copying over extra CA certificates] ********\nSaturday 21 March 2026  00:59:32 +0000 (0:00:01.278)       0:00:47.962 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : neutron | Copying over backend internal TLS certificate] ***\nSaturday 21 March 2026  00:59:37 +0000 (0:00:05.311)       0:00:53.274 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [service-cert-copy : neutron | Copying over backend internal TLS key] *****\nSaturday 21 March 2026  00:59:40 +0000 (0:00:03.103)       0:00:56.378 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Creating TLS backend PEM File] *********************************\nSaturday 21 March 2026  00:59:43 +0000 (0:00:03.223)       0:00:59.602 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Check if policies shall be overwritten] ************************\nSaturday 21 March 2026  00:59:46 +0000 (0:00:02.527)       0:01:02.129 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [neutron : Set neutron policy file] ***************************************\nSaturday 21 March 2026  00:59:46 +0000 (0:00:00.222)       0:01:02.352 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over existing policy file] *****************************\nSaturday 21 March 2026  00:59:46 +0000 (0:00:00.553)       0:01:02.906 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over config.json files for services] *******************\nSaturday 21 March 2026  00:59:50 +0000 (0:00:03.954)       0:01:06.860 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\nTASK [neutron : Copying over neutron.conf] *************************************\nSaturday 21 March 2026  00:59:55 +0000 (0:00:04.097)       0:01:10.958 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\nTASK [neutron : Copying over neutron_vpnaas.conf] ******************************\nSaturday 21 March 2026  01:00:03 +0000 (0:00:08.275)       0:01:19.234 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over ssh key] ******************************************\nSaturday 21 March 2026  01:00:05 +0000 (0:00:02.502)       0:01:21.736 ******** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [neutron : Copying over ml2_conf.ini] *************************************\nSaturday 21 March 2026  01:00:09 +0000 (0:00:03.230)       0:01:24.967 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\nTASK [neutron : Copying over linuxbridge_agent.ini] ****************************\nSaturday 21 March 2026  01:00:13 +0000 (0:00:03.956)       0:01:28.923 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over openvswitch_agent.ini] ****************************\nSaturday 21 March 2026  01:00:15 +0000 (0:00:02.400)       0:01:31.324 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over sriov_agent.ini] **********************************\nSaturday 21 March 2026  01:00:17 +0000 (0:00:02.514)       0:01:33.839 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over mlnx_agent.ini] ***********************************\nSaturday 21 March 2026  01:00:19 +0000 (0:00:01.951)       0:01:35.790 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over eswitchd.conf] ************************************\nSaturday 21 March 2026  01:00:21 +0000 (0:00:02.063)       0:01:37.854 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [neutron : Copying over dhcp_agent.ini] ***********************************\nSaturday 21 March 2026  01:00:23 +0000 (0:00:02.029)       0:01:39.884 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Copying over dnsmasq.conf] *************************************\nSaturday 21 March 2026  01:00:25 +0000 (0:00:01.873)       0:01:41.757 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Copying over l3_agent.ini] *************************************\nSaturday 21 March 2026  01:00:27 +0000 (0:00:01.701)       0:01:43.458 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over fwaas_driver.ini] *********************************\nSaturday 21 March 2026  01:00:29 +0000 (0:00:01.914)       0:01:45.373 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over metadata_agent.ini] *******************************\nSaturday 21 March 2026  01:00:31 +0000 (0:00:01.736)       0:01:47.110 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [neutron : Copying over neutron_ovn_metadata_agent.ini] *******************\nSaturday 21 March 2026  01:00:33 +0000 (0:00:02.401)       0:01:49.511 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over metering_agent.ini] *******************************\nSaturday 21 March 2026  01:00:36 +0000 (0:00:03.075)       0:01:52.587 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over ironic_neutron_agent.ini] *************************\nSaturday 21 March 2026  01:00:38 +0000 (0:00:02.257)       0:01:54.845 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over bgp_dragent.ini] **********************************\nSaturday 21 March 2026  01:00:41 +0000 (0:00:02.655)       0:01:57.500 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over ovn_agent.ini] ************************************\nSaturday 21 March 2026  01:00:44 +0000 (0:00:03.117)       0:02:00.618 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over nsx.ini] ******************************************\nSaturday 21 March 2026  01:00:46 +0000 (0:00:01.630)       0:02:02.248 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copy neutron-l3-agent-wrapper script] **************************\nSaturday 21 March 2026  01:00:49 +0000 (0:00:02.786)       0:02:05.035 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over extra ml2 plugins] ********************************\nSaturday 21 March 2026  01:00:52 +0000 (0:00:02.875)       0:02:07.911 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over neutron-tls-proxy.cfg] ****************************\nSaturday 21 March 2026  01:00:54 +0000 (0:00:02.574)       0:02:10.485 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Copying over neutron_taas.conf] ********************************\nSaturday 21 March 2026  01:00:56 +0000 (0:00:01.909)       0:02:12.395 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Check neutron containers] **************************************\nSaturday 21 March 2026  01:00:58 +0000 (0:00:02.257)       0:02:14.652 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\nTASK [neutron : include_tasks] *************************************************\nSaturday 21 March 2026  01:01:02 +0000 (0:00:03.413)       0:02:18.066 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Creating Neutron database] *************************************\nSaturday 21 March 2026  01:01:02 +0000 (0:00:00.561)       0:02:18.627 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [neutron : Creating Neutron database user and setting permissions] ********\nSaturday 21 March 2026  01:01:04 +0000 (0:00:01.941)       0:02:20.568 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [neutron : Running Neutron bootstrap container] ***************************\nSaturday 21 March 2026  01:01:06 +0000 (0:00:02.320)       0:02:22.888 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [neutron : Flush Handlers] ************************************************\nSaturday 21 March 2026  01:01:43 +0000 (0:00:37.007)       0:02:59.896 ******** \n\nTASK [neutron : Flush Handlers] ************************************************\nSaturday 21 March 2026  01:01:44 +0000 (0:00:00.066)       0:02:59.963 ******** \n\nTASK [neutron : Flush Handlers] ************************************************\nSaturday 21 March 2026  01:01:44 +0000 (0:00:00.112)       0:03:00.075 ******** \n\nTASK [neutron : Flush Handlers] ************************************************\nSaturday 21 March 2026  01:01:44 +0000 (0:00:00.118)       0:03:00.193 ******** \n\nTASK [neutron : Flush Handlers] ************************************************\nSaturday 21 March 2026  01:01:44 +0000 (0:00:00.112)       0:03:00.306 ******** \n\nTASK [neutron : Flush Handlers] ************************************************\nSaturday 21 March 2026  01:01:44 +0000 (0:00:00.070)       0:03:00.377 ******** \n\nRUNNING HANDLER [neutron : Restart neutron-server container] *******************\nSaturday 21 March 2026  01:01:44 +0000 (0:00:00.066)       0:03:00.443 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [neutron : Restart neutron-ovn-metadata-agent container] *******\nSaturday 21 March 2026  01:02:11 +0000 (0:00:26.633)       0:03:27.077 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=31  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=31  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  01:02:53 +0000 (0:00:42.713)       0:04:09.790 ******** \n=============================================================================== \nneutron : Restart neutron-ovn-metadata-agent container ----------------- 42.71s\nneutron : Running Neutron bootstrap container -------------------------- 37.01s\nneutron : Restart neutron-server container ----------------------------- 26.63s\nneutron : Copying over neutron.conf ------------------------------------- 8.28s\nservice-ks-register : neutron | Granting user roles --------------------- 7.30s\nservice-ks-register : neutron | Creating endpoints ---------------------- 6.01s\nservice-cert-copy : neutron | Copying over extra CA certificates -------- 5.31s\nservice-ks-register : neutron | Creating users -------------------------- 4.29s\nneutron : Copying over config.json files for services ------------------- 4.10s\nneutron : Copying over ml2_conf.ini ------------------------------------- 3.96s\nneutron : Copying over existing policy file ----------------------------- 3.95s\nservice-ks-register : neutron | Creating services ----------------------- 3.94s\nneutron : Check neutron containers -------------------------------------- 3.41s\nservice-ks-register : neutron | Creating projects ----------------------- 3.31s\nneutron : Copying over ssh key ------------------------------------------ 3.23s\nservice-ks-register : neutron | Creating roles -------------------------- 3.23s\nservice-cert-copy : neutron | Copying over backend internal TLS key ----- 3.22s\nSetting sysctl values --------------------------------------------------- 3.20s\nneutron : Copying over bgp_dragent.ini ---------------------------------- 3.12s\nservice-cert-copy : neutron | Copying over backend internal TLS certificate --- 3.10s\n\u001b[32m2026-03-21 01:02:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:02:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:02:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:03:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:03:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:03:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:03:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:03:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:03:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:04:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:04:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:04:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:04:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:04:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  01:01:35 +0000 (0:00:00.288)       0:00:00.288 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  01:01:36 +0000 (0:00:00.631)       0:00:00.920 ******** \n\u001b[0;32mok: [testbed-manager] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_prometheus_True)\u001b[0m\n\nPLAY [Apply role prometheus] ***************************************************\n\nTASK [prometheus : include_tasks] **********************************************\nSaturday 21 March 2026  01:01:37 +0000 (0:00:00.765)       0:00:01.685 ******** \n\u001b[0;36mincluded: /ansible/roles/prometheus/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [prometheus : Ensuring config directories exist] **************************\nSaturday 21 March 2026  01:01:38 +0000 (0:00:01.115)       0:00:02.801 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [prometheus : include_tasks] **********************************************\nSaturday 21 March 2026  01:01:41 +0000 (0:00:03.266)       0:00:06.068 ******** \n\u001b[0;36mincluded: /ansible/roles/prometheus/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [service-cert-copy : prometheus | Copying over extra CA certificates] *****\nSaturday 21 March 2026  01:01:43 +0000 (0:00:01.310)       0:00:07.378 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-cert-copy : prometheus | Copying over backend internal TLS certificate] ***\nSaturday 21 March 2026  01:01:49 +0000 (0:00:06.480)       0:00:13.859 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [service-cert-copy : prometheus | Copying over backend internal TLS key] ***\nSaturday 21 March 2026  01:01:51 +0000 (0:00:02.313)       0:00:16.172 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [prometheus : Copying over config.json files] *****************************\nSaturday 21 March 2026  01:01:54 +0000 (0:00:02.707)       0:00:18.880 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [prometheus : Find custom prometheus alert rules files] *******************\nSaturday 21 March 2026  01:02:00 +0000 (0:00:06.364)       0:00:25.244 ******** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [prometheus : Copying over custom prometheus alert rules files] ***********\nSaturday 21 March 2026  01:02:01 +0000 (0:00:01.017)       0:00:26.262 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1106499, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0519824, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1106499, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0519824, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1106499, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0519824, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1106527, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.061685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1106499, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0519824, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1106527, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.061685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1106499, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0519824, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1106527, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.061685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1106493, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.050685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1106493, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.050685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1106499, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0519824, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1106527, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.061685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1106499, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0519824, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1106527, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.061685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1106493, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.050685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1106512, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.055685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1106527, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.061685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1106493, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.050685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1106512, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.055685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1106493, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.050685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1106489, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04872, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1106512, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.055685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1106493, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.050685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1106501, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0524561, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1106512, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.055685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1106527, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.061685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1106489, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04872, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1106512, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.055685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1106489, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04872, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1106512, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.055685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1106510, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0552485, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1106489, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04872, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1106501, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0524561, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1106489, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04872, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1106501, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0524561, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1106503, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.052685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1106489, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04872, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1106501, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0524561, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1106501, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0524561, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1106510, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0552485, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1106493, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.050685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1106510, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0552485, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1106510, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0552485, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1106503, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.052685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1106503, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.052685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1106510, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0552485, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1106501, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0524561, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1106498, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0515304, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1106498, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0515304, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1106503, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.052685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1106498, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0515304, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1106510, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0552485, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1106503, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.052685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106526, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0597665, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1106498, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0515304, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106526, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0597665, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106526, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0597665, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1106498, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0515304, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1106503, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.052685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106526, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0597665, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106486, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0478735, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106526, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0597665, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106486, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0478735, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106486, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0478735, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106486, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0478735, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1106512, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.055685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106486, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0478735, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1106549, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0655851, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1106498, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0515304, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1106549, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0655851, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1106549, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0655851, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1106524, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.058685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1106549, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0655851, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1106549, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0655851, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1106524, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.058685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106526, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0597665, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106490, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0489995, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1106524, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.058685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1106524, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.058685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106486, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0478735, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106490, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0489995, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1106489, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04872, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1106524, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.058685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1106488, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04834, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106490, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0489995, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1106549, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0655851, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1106507, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0546963, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106490, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0489995, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1106488, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04834, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1106488, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04834, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1106505, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0536127, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106490, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0489995, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1106488, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04834, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1106524, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.058685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1106507, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0546963, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1106507, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0546963, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106490, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0489995, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1106545, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0648096, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1106488, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04834, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1106501, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0524561, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1106507, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0546963, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1106505, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0536127, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1106505, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0536127, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1106505, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0536127, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1106488, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04834, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1106507, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0546963, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1106545, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0648096, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1106545, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0648096, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1106545, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0648096, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1106507, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0546963, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1106505, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0536127, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1106510, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0552485, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1106545, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0648096, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1106505, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0536127, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1106545, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0648096, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1106503, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.052685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1106498, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0515304, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106526, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0597665, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106486, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0478735, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1106549, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0655851, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1106524, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.058685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106490, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0489995, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1106488, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04834, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1106507, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0546963, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1106505, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0536127, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1106545, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0648096, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\nTASK [prometheus : Find prometheus common config overrides] ********************\nSaturday 21 March 2026  01:02:27 +0000 (0:00:25.372)       0:00:51.634 ******** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [prometheus : Find prometheus host config overrides] **********************\nSaturday 21 March 2026  01:02:28 +0000 (0:00:00.732)       0:00:52.367 ******** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mmanager/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mmanager/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-0/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-0/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-1/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-1/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-3/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-3/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-2/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-2/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-4/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-4/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-5/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-5/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [prometheus : Copying over prometheus config file] ************************\nSaturday 21 March 2026  01:02:29 +0000 (0:00:01.525)       0:00:53.892 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2)\u001b[0m\n\nTASK [prometheus : Copying over prometheus web config file] ********************\nSaturday 21 March 2026  01:02:43 +0000 (0:00:13.824)       0:01:07.717 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2)\u001b[0m\n\nTASK [prometheus : Copying over prometheus alertmanager config file] ***********\nSaturday 21 March 2026  01:02:47 +0000 (0:00:03.647)       0:01:11.365 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Find custom Alertmanager alert notification templates] ******\nSaturday 21 March 2026  01:02:48 +0000 (0:00:01.640)       0:01:13.005 ******** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [prometheus : Copying over custom Alertmanager alert notification templates] ***\nSaturday 21 March 2026  01:02:49 +0000 (0:00:00.765)       0:01:13.771 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Copying over my.cnf for mysqld_exporter] ********************\nSaturday 21 March 2026  01:02:50 +0000 (0:00:00.869)       0:01:14.640 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [prometheus : Copying cloud config file for openstack exporter] ***********\nSaturday 21 March 2026  01:02:52 +0000 (0:00:02.642)       0:01:17.283 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Copying config file for blackbox exporter] ******************\nSaturday 21 March 2026  01:02:54 +0000 (0:00:01.875)       0:01:19.159 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2)\u001b[0m\n\nTASK [prometheus : Find extra prometheus server config files] ******************\nSaturday 21 March 2026  01:02:56 +0000 (0:00:01.585)       0:01:20.745 ******** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/extras/' path\u001b[0m\n\u001b[1;35mdue to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/extras/' is\u001b[0m\n\u001b[1;35mnot a directory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [prometheus : Create subdirectories for extra config files] ***************\nSaturday 21 March 2026  01:02:57 +0000 (0:00:01.123)       0:01:21.868 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Template extra prometheus server config files] **************\nSaturday 21 March 2026  01:02:58 +0000 (0:00:00.656)       0:01:22.525 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Check prometheus containers] ********************************\nSaturday 21 March 2026  01:02:59 +0000 (0:00:00.976)       0:01:23.502 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [prometheus : Creating prometheus database user and setting permissions] ***\nSaturday 21 March 2026  01:03:03 +0000 (0:00:04.747)       0:01:28.249 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [prometheus : Flush handlers] *********************************************\nSaturday 21 March 2026  01:03:04 +0000 (0:00:00.786)       0:01:29.036 ******** \n\nTASK [prometheus : Flush handlers] *********************************************\nSaturday 21 March 2026  01:03:04 +0000 (0:00:00.052)       0:01:29.088 ******** \n\nTASK [prometheus : Flush handlers] *********************************************\nSaturday 21 March 2026  01:03:04 +0000 (0:00:00.049)       0:01:29.137 ******** \n\nTASK [prometheus : Flush handlers] *********************************************\nSaturday 21 March 2026  01:03:04 +0000 (0:00:00.050)       0:01:29.188 ******** \n\nTASK [prometheus : Flush handlers] *********************************************\nSaturday 21 March 2026  01:03:04 +0000 (0:00:00.050)       0:01:29.238 ******** \n\nTASK [prometheus : Flush handlers] *********************************************\nSaturday 21 March 2026  01:03:04 +0000 (0:00:00.050)       0:01:29.289 ******** \n\nTASK [prometheus : Flush handlers] *********************************************\nSaturday 21 March 2026  01:03:05 +0000 (0:00:00.051)       0:01:29.340 ******** \n\nRUNNING HANDLER [prometheus : Restart prometheus-server container] *************\nSaturday 21 March 2026  01:03:05 +0000 (0:00:00.069)       0:01:29.410 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-node-exporter container] ******\nSaturday 21 March 2026  01:03:20 +0000 (0:00:15.705)       0:01:45.115 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-mysqld-exporter container] ****\nSaturday 21 March 2026  01:03:33 +0000 (0:00:12.888)       0:01:58.004 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-memcached-exporter container] ***\nSaturday 21 March 2026  01:03:43 +0000 (0:00:09.801)       0:02:07.806 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-cadvisor container] ***********\nSaturday 21 March 2026  01:03:49 +0000 (0:00:05.972)       0:02:13.778 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-alertmanager container] *******\nSaturday 21 March 2026  01:04:04 +0000 (0:00:15.348)       0:02:29.126 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-elasticsearch-exporter container] ***\nSaturday 21 March 2026  01:04:11 +0000 (0:00:06.529)       0:02:35.656 ******** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-blackbox-exporter container] ***\nSaturday 21 March 2026  01:04:20 +0000 (0:00:09.515)       0:02:45.171 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-libvirt-exporter container] ***\nSaturday 21 March 2026  01:04:30 +0000 (0:00:09.884)       0:02:55.056 ******** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  01:04:39 +0000 (0:00:08.892)       0:03:03.948 ******** \n=============================================================================== \nprometheus : Copying over custom prometheus alert rules files ---------- 25.37s\nprometheus : Restart prometheus-server container ----------------------- 15.71s\nprometheus : Restart prometheus-cadvisor container --------------------- 15.35s\nprometheus : Copying over prometheus config file ----------------------- 13.82s\nprometheus : Restart prometheus-node-exporter container ---------------- 12.89s\nprometheus : Restart prometheus-blackbox-exporter container ------------- 9.88s\nprometheus : Restart prometheus-mysqld-exporter container --------------- 9.80s\nprometheus : Restart prometheus-elasticsearch-exporter container -------- 9.52s\nprometheus : Restart prometheus-libvirt-exporter container -------------- 8.89s\nprometheus : Restart prometheus-alertmanager container ------------------ 6.53s\nservice-cert-copy : prometheus | Copying over extra CA certificates ----- 6.48s\nprometheus : Copying over config.json files ----------------------------- 6.36s\nprometheus : Restart prometheus-memcached-exporter container ------------ 5.97s\nprometheus : Check prometheus containers -------------------------------- 4.75s\nprometheus : Copying over prometheus web config file -------------------- 3.65s\nprometheus : Ensuring config directories exist -------------------------- 3.27s\nservice-cert-copy : prometheus | Copying over backend internal TLS key --- 2.71s\nprometheus : Copying over my.cnf for mysqld_exporter -------------------- 2.64s\nservice-cert-copy : prometheus | Copying over backend internal TLS certificate --- 2.31s\nprometheus : Copying cloud config file for openstack exporter ----------- 1.88s\n\u001b[32m2026-03-21 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:04:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:04:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:04:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:04:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:04:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:04:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:05:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:05:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  01:02:20 +0000 (0:00:00.312)       0:00:00.312 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  01:02:21 +0000 (0:00:00.265)       0:00:00.578 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_glance_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_glance_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_glance_True)\u001b[0m\n\nPLAY [Apply role glance] *******************************************************\n\nTASK [glance : include_tasks] **************************************************\nSaturday 21 March 2026  01:02:21 +0000 (0:00:00.258)       0:00:00.836 ******** \n\u001b[0;36mincluded: /ansible/roles/glance/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : glance | Creating services] ************************\nSaturday 21 March 2026  01:02:22 +0000 (0:00:00.915)       0:00:01.751 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=glance (image))\u001b[0m\n\nTASK [service-ks-register : glance | Creating endpoints] ***********************\nSaturday 21 March 2026  01:02:25 +0000 (0:00:03.327)       0:00:05.079 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> https://api-int.testbed.osism.xyz:9292 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> https://api.testbed.osism.xyz:9292 -> public)\u001b[0m\n\nTASK [service-ks-register : glance | Creating projects] ************************\nSaturday 21 March 2026  01:02:31 +0000 (0:00:05.636)       0:00:10.715 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : glance | Creating users] ***************************\nSaturday 21 March 2026  01:02:33 +0000 (0:00:02.671)       0:00:13.387 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : glance | Creating roles] ***************************\nSaturday 21 March 2026  01:02:37 +0000 (0:00:03.388)       0:00:16.775 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : glance | Granting user roles] **********************\nSaturday 21 March 2026  01:02:40 +0000 (0:00:03.066)       0:00:19.841 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> service -> admin)\u001b[0m\n\nTASK [glance : Ensuring config directories exist] ******************************\nSaturday 21 March 2026  01:02:44 +0000 (0:00:03.645)       0:00:23.487 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\nTASK [glance : include_tasks] **************************************************\nSaturday 21 March 2026  01:02:48 +0000 (0:00:04.504)       0:00:27.994 ******** \n\u001b[0;36mincluded: /ansible/roles/glance/tasks/external_ceph.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [glance : Ensuring glance service ceph config subdir exists] **************\nSaturday 21 March 2026  01:02:49 +0000 (0:00:00.580)       0:00:28.574 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [glance : Copy over multiple ceph configs for Glance] *********************\nSaturday 21 March 2026  01:02:53 +0000 (0:00:04.261)       0:00:32.836 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\nTASK [glance : Copy over ceph Glance keyrings] *********************************\nSaturday 21 March 2026  01:02:55 +0000 (0:00:02.022)       0:00:34.858 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\nTASK [glance : Ensuring config directory has correct owner and permission] *****\nSaturday 21 March 2026  01:02:56 +0000 (0:00:01.542)       0:00:36.401 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [glance : Check if policies shall be overwritten] *************************\nSaturday 21 March 2026  01:02:57 +0000 (0:00:00.669)       0:00:37.071 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [glance : Set glance policy file] *****************************************\nSaturday 21 March 2026  01:02:57 +0000 (0:00:00.097)       0:00:37.168 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : include_tasks] **************************************************\nSaturday 21 March 2026  01:02:58 +0000 (0:00:00.313)       0:00:37.482 ******** \n\u001b[0;36mincluded: /ansible/roles/glance/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : glance | Copying over extra CA certificates] *********\nSaturday 21 March 2026  01:02:58 +0000 (0:00:00.586)       0:00:38.069 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\nTASK [service-cert-copy : glance | Copying over backend internal TLS certificate] ***\nSaturday 21 March 2026  01:03:03 +0000 (0:00:04.907)       0:00:42.977 ******** \n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [service-cert-copy : glance | Copying over backend internal TLS key] ******\nSaturday 21 March 2026  01:03:06 +0000 (0:00:02.683)       0:00:45.660 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Creating TLS backend PEM File] **********************************\nSaturday 21 March 2026  01:03:09 +0000 (0:00:03.125)       0:00:48.786 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [glance : Copying over config.json files for services] ********************\nSaturday 21 March 2026  01:03:12 +0000 (0:00:03.485)       0:00:52.272 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\nTASK [glance : Copying over glance-api.conf] ***********************************\nSaturday 21 March 2026  01:03:17 +0000 (0:00:04.225)       0:00:56.497 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [glance : Copying over glance-cache.conf for glance_api] ******************\nSaturday 21 March 2026  01:03:25 +0000 (0:00:08.101)       0:01:04.599 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [glance : Copying over glance-swift.conf for glance_api] ******************\nSaturday 21 March 2026  01:03:29 +0000 (0:00:04.519)       0:01:09.118 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Copying over glance-image-import.conf] **************************\nSaturday 21 March 2026  01:03:33 +0000 (0:00:03.360)       0:01:12.479 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Copying over property-protections-rules.conf] *******************\nSaturday 21 March 2026  01:03:36 +0000 (0:00:03.940)       0:01:16.420 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Copying over existing policy file] ******************************\nSaturday 21 March 2026  01:03:39 +0000 (0:00:02.907)       0:01:19.328 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Copying over glance-haproxy-tls.cfg] ****************************\nSaturday 21 March 2026  01:03:40 +0000 (0:00:00.371)       0:01:19.699 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Generating 'hostnqn' file for glance_api] ***********************\nSaturday 21 March 2026  01:03:43 +0000 (0:00:03.511)       0:01:23.211 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Generating 'hostid' file for glance_api] ************************\nSaturday 21 March 2026  01:03:48 +0000 (0:00:04.657)       0:01:27.868 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Check glance containers] ****************************************\nSaturday 21 March 2026  01:03:54 +0000 (0:00:06.562)       0:01:34.430 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\nTASK [glance : include_tasks] **************************************************\nSaturday 21 March 2026  01:04:00 +0000 (0:00:05.160)       0:01:39.591 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Creating Glance database] ***************************************\nSaturday 21 March 2026  01:04:00 +0000 (0:00:00.443)       0:01:40.034 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Creating Glance database user and setting permissions] **********\nSaturday 21 March 2026  01:04:02 +0000 (0:00:01.860)       0:01:41.894 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Enable log_bin_trust_function_creators function] ****************\nSaturday 21 March 2026  01:04:04 +0000 (0:00:01.834)       0:01:43.729 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Running Glance bootstrap container] *****************************\nSaturday 21 March 2026  01:04:05 +0000 (0:00:01.731)       0:01:45.460 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Disable log_bin_trust_function_creators function] ***************\nSaturday 21 March 2026  01:04:32 +0000 (0:00:26.793)       0:02:12.254 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Flush handlers] *************************************************\nSaturday 21 March 2026  01:04:34 +0000 (0:00:01.715)       0:02:13.970 ******** \n\nTASK [glance : Flush handlers] *************************************************\nSaturday 21 March 2026  01:04:34 +0000 (0:00:00.064)       0:02:14.034 ******** \n\nTASK [glance : Flush handlers] *************************************************\nSaturday 21 March 2026  01:04:34 +0000 (0:00:00.061)       0:02:14.096 ******** \n\nRUNNING HANDLER [glance : Restart glance-api container] ************************\nSaturday 21 March 2026  01:04:34 +0000 (0:00:00.065)       0:02:14.162 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=14  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=13  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=13  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  01:05:01 +0000 (0:00:26.327)       0:02:40.490 ******** \n=============================================================================== \nglance : Running Glance bootstrap container ---------------------------- 26.79s\nglance : Restart glance-api container ---------------------------------- 26.33s\nglance : Copying over glance-api.conf ----------------------------------- 8.10s\nglance : Generating 'hostid' file for glance_api ------------------------ 6.56s\nservice-ks-register : glance | Creating endpoints ----------------------- 5.64s\nglance : Check glance containers ---------------------------------------- 5.16s\nservice-cert-copy : glance | Copying over extra CA certificates --------- 4.91s\nglance : Generating 'hostnqn' file for glance_api ----------------------- 4.66s\nglance : Copying over glance-cache.conf for glance_api ------------------ 4.52s\nglance : Ensuring config directories exist ------------------------------ 4.50s\nglance : Ensuring glance service ceph config subdir exists -------------- 4.26s\nglance : Copying over config.json files for services -------------------- 4.23s\nglance : Copying over glance-image-import.conf -------------------------- 3.94s\nservice-ks-register : glance | Granting user roles ---------------------- 3.65s\nglance : Copying over glance-haproxy-tls.cfg ---------------------------- 3.51s\nglance : Creating TLS backend PEM File ---------------------------------- 3.49s\nservice-ks-register : glance | Creating users --------------------------- 3.39s\nglance : Copying over glance-swift.conf for glance_api ------------------ 3.36s\nservice-ks-register : glance | Creating services ------------------------ 3.33s\nservice-cert-copy : glance | Copying over backend internal TLS key ------ 3.13s\n\u001b[32m2026-03-21 01:05:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:05:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:05:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:05:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:05:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:05:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  01:02:49 +0000 (0:00:00.269)       0:00:00.269 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  01:02:50 +0000 (0:00:00.392)       0:00:00.661 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_cinder_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_cinder_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_cinder_True)\u001b[0m\n\nPLAY [Apply role cinder] *******************************************************\n\nTASK [cinder : include_tasks] **************************************************\nSaturday 21 March 2026  01:02:50 +0000 (0:00:00.287)       0:00:00.949 ******** \n\u001b[0;36mincluded: /ansible/roles/cinder/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : cinder | Creating services] ************************\nSaturday 21 March 2026  01:02:51 +0000 (0:00:00.870)       0:00:01.819 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 (volumev3))\u001b[0m\n\nTASK [service-ks-register : cinder | Creating endpoints] ***********************\nSaturday 21 March 2026  01:02:54 +0000 (0:00:03.745)       0:00:05.564 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 -> https://api-int.testbed.osism.xyz:8776/v3/%(tenant_id)s -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 -> https://api.testbed.osism.xyz:8776/v3/%(tenant_id)s -> public)\u001b[0m\n\nTASK [service-ks-register : cinder | Creating projects] ************************\nSaturday 21 March 2026  01:03:00 +0000 (0:00:05.802)       0:00:11.366 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : cinder | Creating users] ***************************\nSaturday 21 March 2026  01:03:04 +0000 (0:00:03.287)       0:00:14.654 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : cinder | Creating roles] ***************************\nSaturday 21 March 2026  01:03:07 +0000 (0:00:03.610)       0:00:18.264 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : cinder | Granting user roles] **********************\nSaturday 21 March 2026  01:03:10 +0000 (0:00:03.122)       0:00:21.387 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service -> admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service -> service)\u001b[0m\n\nTASK [cinder : Ensuring config directories exist] ******************************\nSaturday 21 March 2026  01:03:17 +0000 (0:00:07.115)       0:00:28.502 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [cinder : include_tasks] **************************************************\nSaturday 21 March 2026  01:03:20 +0000 (0:00:02.931)       0:00:31.433 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : include_tasks] **************************************************\nSaturday 21 March 2026  01:03:21 +0000 (0:00:00.555)       0:00:31.989 ******** \n\u001b[0;36mincluded: /ansible/roles/cinder/tasks/external_ceph.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [cinder : Ensuring cinder service ceph config subdirs exists] *************\nSaturday 21 March 2026  01:03:23 +0000 (0:00:01.670)       0:00:33.659 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder-volume)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=cinder-volume)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=cinder-volume)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder-backup)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=cinder-backup)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=cinder-backup)\u001b[0m\n\nTASK [cinder : Copying over multiple ceph.conf for cinder services] ************\nSaturday 21 March 2026  01:03:26 +0000 (0:00:03.337)       0:00:36.996 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m\n\nTASK [cinder : Copy over Ceph keyring files for cinder-volume] *****************\nSaturday 21 March 2026  01:03:30 +0000 (0:00:03.928)       0:00:40.924 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\nTASK [cinder : Copy over Ceph keyring files for cinder-backup] *****************\nSaturday 21 March 2026  01:03:31 +0000 (0:00:01.667)       0:00:42.592 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\nTASK [cinder : Ensuring config directory has correct owner and permission] *****\nSaturday 21 March 2026  01:03:34 +0000 (0:00:02.751)       0:00:45.344 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=cinder-volume)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=cinder-volume)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=cinder-backup)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=cinder-volume)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=cinder-backup)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=cinder-backup)\u001b[0m\n\nTASK [cinder : Check if policies shall be overwritten] *************************\nSaturday 21 March 2026  01:03:35 +0000 (0:00:01.291)       0:00:46.635 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [cinder : Set cinder policy file] *****************************************\nSaturday 21 March 2026  01:03:36 +0000 (0:00:00.160)       0:00:46.796 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : include_tasks] **************************************************\nSaturday 21 March 2026  01:03:36 +0000 (0:00:00.415)       0:00:47.211 ******** \n\u001b[0;36mincluded: /ansible/roles/cinder/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : cinder | Copying over extra CA certificates] *********\nSaturday 21 March 2026  01:03:37 +0000 (0:00:00.512)       0:00:47.724 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : cinder | Copying over backend internal TLS certificate] ***\nSaturday 21 March 2026  01:03:41 +0000 (0:00:03.992)       0:00:51.717 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [service-cert-copy : cinder | Copying over backend internal TLS key] ******\nSaturday 21 March 2026  01:03:41 +0000 (0:00:00.835)       0:00:52.553 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : Copying over config.json files for services] ********************\nSaturday 21 March 2026  01:03:42 +0000 (0:00:01.096)       0:00:53.649 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [cinder : Copying over cinder-wsgi.conf] **********************************\nSaturday 21 March 2026  01:03:47 +0000 (0:00:04.254)       0:00:57.904 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)\u001b[0m\n\nTASK [cinder : Copying over cinder.conf] ***************************************\nSaturday 21 March 2026  01:03:49 +0000 (0:00:01.808)       0:00:59.712 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [cinder : Generating 'hostnqn' file for cinder_volume] ********************\nSaturday 21 March 2026  01:04:03 +0000 (0:00:14.625)       0:01:14.338 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [cinder : Generating 'hostid' file for cinder_volume] *********************\nSaturday 21 March 2026  01:04:04 +0000 (0:00:01.276)       0:01:15.614 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [cinder : Copying over existing policy file] ******************************\nSaturday 21 March 2026  01:04:06 +0000 (0:00:01.214)       0:01:16.829 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : Copying over nfs_shares files for cinder_volume] ****************\nSaturday 21 March 2026  01:04:07 +0000 (0:00:00.872)       0:01:17.701 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : Check cinder containers] ****************************************\nSaturday 21 March 2026  01:04:07 +0000 (0:00:00.288)       0:01:17.990 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [cinder : include_tasks] **************************************************\nSaturday 21 March 2026  01:04:10 +0000 (0:00:02.774)       0:01:20.764 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : Creating Cinder database] ***************************************\nSaturday 21 March 2026  01:04:10 +0000 (0:00:00.352)       0:01:21.116 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [cinder : Creating Cinder database user and setting permissions] **********\nSaturday 21 March 2026  01:04:12 +0000 (0:00:02.065)       0:01:23.182 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [cinder : Running Cinder bootstrap container] *****************************\nSaturday 21 March 2026  01:04:14 +0000 (0:00:02.321)       0:01:25.504 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [cinder : Flush handlers] *************************************************\nSaturday 21 March 2026  01:04:32 +0000 (0:00:17.568)       0:01:43.073 ******** \n\nTASK [cinder : Flush handlers] *************************************************\nSaturday 21 March 2026  01:04:32 +0000 (0:00:00.060)       0:01:43.133 ******** \n\nTASK [cinder : Flush handlers] *************************************************\nSaturday 21 March 2026  01:04:32 +0000 (0:00:00.060)       0:01:43.194 ******** \n\nRUNNING HANDLER [cinder : Restart cinder-api container] ************************\nSaturday 21 March 2026  01:04:32 +0000 (0:00:00.060)       0:01:43.254 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [cinder : Restart cinder-scheduler container] ******************\nSaturday 21 March 2026  01:04:48 +0000 (0:00:15.493)       0:01:58.748 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [cinder : Restart cinder-volume container] *********************\nSaturday 21 March 2026  01:04:57 +0000 (0:00:09.775)       0:02:08.524 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [cinder : Restart cinder-backup container] *********************\nSaturday 21 March 2026  01:05:14 +0000 (0:00:16.801)       0:02:25.325 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [cinder : Wait for cinder services to update service versions] ***\nSaturday 21 March 2026  01:05:20 +0000 (0:00:05.377)       0:02:30.703 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=31  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  01:05:20 +0000 (0:00:00.273)       0:02:30.976 ******** \n=============================================================================== \ncinder : Running Cinder bootstrap container ---------------------------- 17.57s\ncinder : Restart cinder-volume container ------------------------------- 16.80s\ncinder : Restart cinder-api container ---------------------------------- 15.49s\ncinder : Copying over cinder.conf -------------------------------------- 14.63s\ncinder : Restart cinder-scheduler container ----------------------------- 9.78s\nservice-ks-register : cinder | Granting user roles ---------------------- 7.12s\nservice-ks-register : cinder | Creating endpoints ----------------------- 5.81s\ncinder : Restart cinder-backup container -------------------------------- 5.38s\ncinder : Copying over config.json files for services -------------------- 4.25s\nservice-cert-copy : cinder | Copying over extra CA certificates --------- 3.99s\ncinder : Copying over multiple ceph.conf for cinder services ------------ 3.93s\nservice-ks-register : cinder | Creating services ------------------------ 3.75s\nservice-ks-register : cinder | Creating users --------------------------- 3.61s\ncinder : Ensuring cinder service ceph config subdirs exists ------------- 3.34s\nservice-ks-register : cinder | Creating projects ------------------------ 3.28s\nservice-ks-register : cinder | Creating roles --------------------------- 3.12s\ncinder : Ensuring config directories exist ------------------------------ 2.93s\ncinder : Check cinder containers ---------------------------------------- 2.77s\ncinder : Copy over Ceph keyring files for cinder-backup ----------------- 2.75s\ncinder : Creating Cinder database user and setting permissions ---------- 2.32s\n\u001b[32m2026-03-21 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:05:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:05:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:05:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:05:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:05:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:05:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:05:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:05:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:06:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:06:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:06:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:06:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:06:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:06:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:06:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:06:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:06:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-21 01:06:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:06:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  01:04:43 +0000 (0:00:00.222)       0:00:00.222 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  01:04:43 +0000 (0:00:00.372)       0:00:00.595 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_nova_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_nova_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_nova_True)\u001b[0m\n\nPLAY [Wait for the Nova service] ***********************************************\n\nTASK [Waiting for Nova public port to be UP] ***********************************\nSaturday 21 March 2026  01:04:44 +0000 (0:00:00.521)       0:00:01.116 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  01:06:39 +0000 (0:01:55.136)       0:01:56.253 ******** \n=============================================================================== \nWaiting for Nova public port to be UP --------------------------------- 115.13s\nGroup hosts based on enabled services ----------------------------------- 0.52s\nGroup hosts based on Kolla action --------------------------------------- 0.37s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  01:05:04 +0000 (0:00:00.309)       0:00:00.309 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  01:05:04 +0000 (0:00:00.284)       0:00:00.593 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_grafana_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_grafana_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_grafana_True)\u001b[0m\n\nPLAY [Apply role grafana] ******************************************************\n\nTASK [grafana : include_tasks] *************************************************\nSaturday 21 March 2026  01:05:05 +0000 (0:00:00.303)       0:00:00.897 ******** \n\u001b[0;36mincluded: /ansible/roles/grafana/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [grafana : Ensuring config directories exist] *****************************\nSaturday 21 March 2026  01:05:05 +0000 (0:00:00.491)       0:00:01.389 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\nTASK [grafana : Check if extra configuration file exists] **********************\nSaturday 21 March 2026  01:05:06 +0000 (0:00:01.157)       0:00:02.547 ******** \n\u001b[1;35m[WARNING]: Skipped '/operations/prometheus/grafana' path due to this access\u001b[0m\n\u001b[1;35missue: '/operations/prometheus/grafana' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [grafana : include_tasks] *************************************************\nSaturday 21 March 2026  01:05:07 +0000 (0:00:00.984)       0:00:03.531 ******** \n\u001b[0;36mincluded: /ansible/roles/grafana/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : grafana | Copying over extra CA certificates] ********\nSaturday 21 March 2026  01:05:08 +0000 (0:00:00.507)       0:00:04.038 ******** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\nTASK [service-cert-copy : grafana | Copying over backend internal TLS certificate] ***\nSaturday 21 March 2026  01:05:09 +0000 (0:00:01.330)       0:00:05.369 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : grafana | Copying over backend internal TLS key] *****\nSaturday 21 March 2026  01:05:09 +0000 (0:00:00.372)       0:00:05.742 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Copying over config.json files] ********************************\nSaturday 21 March 2026  01:05:10 +0000 (0:00:00.577)       0:00:06.319 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\nTASK [grafana : Copying over grafana.ini] **************************************\nSaturday 21 March 2026  01:05:11 +0000 (0:00:01.314)       0:00:07.634 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\nTASK [grafana : Copying over extra configuration file] *************************\nSaturday 21 March 2026  01:05:13 +0000 (0:00:01.189)       0:00:08.823 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Configuring Prometheus as data source for Grafana] *************\nSaturday 21 March 2026  01:05:13 +0000 (0:00:00.301)       0:00:09.125 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m\n\nTASK [grafana : Configuring dashboards provisioning] ***************************\nSaturday 21 March 2026  01:05:14 +0000 (0:00:01.086)       0:00:10.212 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m\n\nTASK [grafana : Find custom grafana dashboards] ********************************\nSaturday 21 March 2026  01:05:15 +0000 (0:00:01.061)       0:00:11.273 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [grafana : Find templated grafana dashboards] *****************************\nSaturday 21 March 2026  01:05:16 +0000 (0:00:01.041)       0:00:12.315 ******** \n\u001b[1;35m[WARNING]: Skipped '/etc/kolla/grafana/dashboards' path due to this access\u001b[0m\n\u001b[1;35missue: '/etc/kolla/grafana/dashboards' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Prune templated Grafana dashboards] ****************************\nSaturday 21 March 2026  01:05:17 +0000 (0:00:00.657)       0:00:12.972 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Copying over custom dashboards] ********************************\nSaturday 21 March 2026  01:05:17 +0000 (0:00:00.304)       0:00:13.277 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 121701, 'inode': 1106291, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9733195, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 121701, 'inode': 1106291, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9733195, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 121701, 'inode': 1106291, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9733195, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/cephfsdashboard.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfsdashboard.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 143913, 'inode': 1106321, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.985684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/cephfsdashboard.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfsdashboard.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 143913, 'inode': 1106321, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.985684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/cephfsdashboard.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfsdashboard.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 143913, 'inode': 1106321, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.985684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26019, 'inode': 1106342, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9952614, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26019, 'inode': 1106342, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9952614, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26019, 'inode': 1106342, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9952614, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1106319, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.984684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1106319, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.984684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1106319, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.984684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 170293, 'inode': 1106343, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.996892, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 170293, 'inode': 1106343, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.996892, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 170293, 'inode': 1106343, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.996892, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph-nvmeof-performance.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-nvmeof-performance.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 33297, 'inode': 1106305, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.978684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph-nvmeof-performance.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-nvmeof-performance.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 33297, 'inode': 1106305, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.978684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph-nvmeof-performance.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-nvmeof-performance.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 33297, 'inode': 1106305, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.978684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26346, 'inode': 1106330, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.989684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26346, 'inode': 1106330, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.989684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26346, 'inode': 1106330, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.989684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 46110, 'inode': 1106339, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9926841, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 46110, 'inode': 1106339, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9926841, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 46110, 'inode': 1106339, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9926841, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1106287, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.971438, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1106287, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.971438, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1106287, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.971438, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1106300, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9748197, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1106300, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9748197, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1106300, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9748197, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1106320, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.984684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1106320, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.984684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1106320, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.984684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19231, 'inode': 1106334, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9914844, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19231, 'inode': 1106334, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9914844, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19231, 'inode': 1106334, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9914844, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13320, 'inode': 1106341, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9948537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13320, 'inode': 1106341, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9948537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13320, 'inode': 1106341, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9948537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1106316, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.982684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1106316, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.982684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1106316, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.982684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 20042, 'inode': 1106338, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9926841, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 20042, 'inode': 1106338, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9926841, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 20042, 'inode': 1106338, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9926841, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/smb-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/smb-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29877, 'inode': 1106347, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9976842, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/smb-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/smb-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29877, 'inode': 1106347, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9976842, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/smb-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/smb-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29877, 'inode': 1106347, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9976842, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38375, 'inode': 1106331, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.990684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38375, 'inode': 1106331, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.990684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38375, 'inode': 1106331, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.990684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 63043, 'inode': 1106328, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9889889, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 63043, 'inode': 1106328, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9889889, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 63043, 'inode': 1106328, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9889889, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27387, 'inode': 1106324, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9882722, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27387, 'inode': 1106324, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9882722, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27387, 'inode': 1106324, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9882722, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49016, 'inode': 1106336, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.991684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49016, 'inode': 1106336, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.991684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49016, 'inode': 1106336, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.991684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 43303, 'inode': 1106322, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.986684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 43303, 'inode': 1106322, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.986684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 43303, 'inode': 1106322, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.986684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16614, 'inode': 1106340, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.993684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16614, 'inode': 1106340, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.993684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16614, 'inode': 1106340, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.993684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph-nvmeof.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-nvmeof.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 52667, 'inode': 1106314, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9806838, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph-nvmeof.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-nvmeof.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 52667, 'inode': 1106314, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9806838, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph-nvmeof.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-nvmeof.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 52667, 'inode': 1106314, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9806838, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1106481, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.044685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1106481, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.044685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1106481, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.044685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1106414, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0217268, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1106414, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0217268, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1106414, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0217268, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1106393, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.014711, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1106393, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.014711, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1106393, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.014711, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15767, 'inode': 1106442, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0248306, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15767, 'inode': 1106442, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0248306, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15767, 'inode': 1106442, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0248306, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1106350, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9986842, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1106350, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9986842, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1106350, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9986842, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1106461, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0348463, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1106461, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0348463, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1106461, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0348463, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1106444, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0326846, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1106444, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0326846, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1106444, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0326846, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22303, 'inode': 1106462, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0355852, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22303, 'inode': 1106462, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0355852, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22303, 'inode': 1106462, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0355852, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1106476, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0416849, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1106476, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0416849, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1106476, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0416849, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21194, 'inode': 1106459, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0336847, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21194, 'inode': 1106459, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0336847, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21194, 'inode': 1106459, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0336847, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1106432, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0229754, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1106432, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0229754, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1106432, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0229754, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1106409, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0184097, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1106409, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0184097, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1106409, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0184097, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1106426, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0226433, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1106426, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0226433, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1106426, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0226433, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1106398, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0175004, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1106398, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0175004, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1106398, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0175004, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15957, 'inode': 1106437, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0236845, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15957, 'inode': 1106437, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0236845, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15957, 'inode': 1106437, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0236845, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1106469, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0406847, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1106469, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0406847, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1106469, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0406847, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1106466, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0376847, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1106466, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0376847, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1106466, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0376847, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1106351, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9995236, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1106351, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9995236, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1106351, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9995236, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1106352, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0139942, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1106352, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0139942, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1106352, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0139942, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1106457, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0336847, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1106457, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0336847, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1106457, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0336847, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21951, 'inode': 1106464, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0366848, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21951, 'inode': 1106464, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0366848, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21951, 'inode': 1106464, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0366848, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\nTASK [grafana : Check grafana containers] **************************************\nSaturday 21 March 2026  01:05:52 +0000 (0:00:34.612)       0:00:47.889 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\nTASK [grafana : Creating grafana database] *************************************\nSaturday 21 March 2026  01:05:53 +0000 (0:00:00.992)       0:00:48.882 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [grafana : Creating grafana database user and setting permissions] ********\nSaturday 21 March 2026  01:05:55 +0000 (0:00:01.902)       0:00:50.784 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [grafana : Flush handlers] ************************************************\nSaturday 21 March 2026  01:05:56 +0000 (0:00:01.794)       0:00:52.579 ******** \n\nTASK [grafana : Flush handlers] ************************************************\nSaturday 21 March 2026  01:05:56 +0000 (0:00:00.059)       0:00:52.639 ******** \n\nTASK [grafana : Flush handlers] ************************************************\nSaturday 21 March 2026  01:05:56 +0000 (0:00:00.064)       0:00:52.703 ******** \n\nRUNNING HANDLER [grafana : Restart first grafana container] ********************\nSaturday 21 March 2026  01:05:56 +0000 (0:00:00.065)       0:00:52.768 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [grafana : Waiting for grafana to start on first node] *********\nSaturday 21 March 2026  01:06:03 +0000 (0:00:06.619)       0:00:59.388 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (12 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [grafana : Restart remaining grafana containers] ***************\nSaturday 21 March 2026  01:06:17 +0000 (0:00:13.744)       0:01:13.133 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [grafana : Wait for grafana application ready] ****************************\nSaturday 21 March 2026  01:06:37 +0000 (0:00:19.766)       0:01:32.900 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [grafana : Remove old grafana docker volume] ******************************\nSaturday 21 March 2026  01:06:39 +0000 (0:00:01.914)       0:01:34.814 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Enable grafana datasources] ************************************\nSaturday 21 March 2026  01:06:39 +0000 (0:00:00.263)       0:01:35.077 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'influxdb', 'value': {'enabled': False, 'data': {'isDefault': True, 'database': 'telegraf', 'name': 'telegraf', 'type': 'influxdb', 'url': 'https://api-int.testbed.osism.xyz:8086', 'access': 'proxy', 'basicAuth': False}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'enabled': True, 'data': {'name': 'opensearch', 'type': 'grafana-opensearch-datasource', 'access': 'proxy', 'url': 'https://api-int.testbed.osism.xyz:9200', 'jsonData': {'flavor': 'OpenSearch', 'database': 'flog-*', 'version': '2.11.1', 'timeField': '@timestamp', 'logLevelField': 'log_level'}}}})\u001b[0m\n\nTASK [grafana : Disable Getting Started panel] *********************************\nSaturday 21 March 2026  01:06:41 +0000 (0:00:02.287)       0:01:37.364 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  01:06:41 +0000 (0:00:00.262)       0:01:37.626 ******** \n=============================================================================== \ngrafana : Copying over custom dashboards ------------------------------- 34.61s\ngrafana : Restart remaining grafana containers ------------------------- 19.77s\ngrafana : Waiting for grafana to start on first node ------------------- 13.74s\ngrafana : Restart first grafana container ------------------------------- 6.62s\ngrafana : Enable grafana datasources ------------------------------------ 2.29s\ngrafana : Wait for grafana application ready ---------------------------- 1.91s\ngrafana : Creating grafana database ------------------------------------- 1.90s\ngrafana : Creating grafana database user and setting permissions -------- 1.79s\nservice-cert-copy : grafana | Copying over extra CA certificates -------- 1.33s\ngrafana : Copying over config.json files -------------------------------- 1.31s\ngrafana : Copying over grafana.ini -------------------------------------- 1.19s\ngrafana : Ensuring config directories exist ----------------------------- 1.16s\ngrafana : Configuring Prometheus as data source for Grafana ------------- 1.09s\ngrafana : Configuring dashboards provisioning --------------------------- 1.06s\ngrafana : Find custom grafana dashboards -------------------------------- 1.04s\ngrafana : Check grafana containers -------------------------------------- 0.99s\ngrafana : Check if extra configuration file exists ---------------------- 0.98s\ngrafana : Find templated grafana dashboards ----------------------------- 0.66s\nservice-cert-copy : grafana | Copying over backend internal TLS key ----- 0.58s\ngrafana : include_tasks ------------------------------------------------- 0.51s\n\u001b[32m2026-03-21 01:06:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:06:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:06:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:06:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:06:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:06:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:06:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:07:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:07:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:07:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:07:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:07:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:07:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:07:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:07:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:07:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:07:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:07:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:07:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:07:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:07:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:07:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:07:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:07:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:07:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:07:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:07:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:08:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:08:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:08:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:08:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:08:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:08:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:08:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:08:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:08:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:08:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:08:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:08:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:08:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:08:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:08:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:08:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:08:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:08:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:08:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:08:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:08:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:09:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:09:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:09:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:09:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:09:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:09:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:09:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:09:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:09:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:09:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:09:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:09:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:09:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:09:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:09:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:09:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:09:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:09:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:09:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:09:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:09:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:10:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:10:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:10:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:10:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:10:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:10:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:10:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:10:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:10:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:10:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:10:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:10:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:10:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:10:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:10:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on OpenStack release] **********************************\nSaturday 21 March 2026  01:02:57 +0000 (0:00:00.278)       0:00:00.278 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  01:02:58 +0000 (0:00:00.534)       0:00:00.812 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  01:02:59 +0000 (0:00:00.736)       0:00:01.548 ******** \n\u001b[0;33mchanged: [testbed-manager] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=enable_nova_True)\u001b[0m\n\nPLAY [Bootstrap nova API databases] ********************************************\n\nTASK [Bootstrap deploy] ********************************************************\nSaturday 21 March 2026  01:03:00 +0000 (0:00:01.137)       0:00:02.686 ******** \n\u001b[0;36mincluded: nova for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova : Creating Nova databases] ******************************************\nSaturday 21 March 2026  01:03:01 +0000 (0:00:01.009)       0:00:03.696 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=nova_cell0)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova_api)\u001b[0m\n\nTASK [nova : Creating Nova databases user and setting permissions] *************\nSaturday 21 March 2026  01:03:05 +0000 (0:00:04.346)       0:00:08.042 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova : Ensuring config directories exist] ********************************\nSaturday 21 March 2026  01:03:09 +0000 (0:00:04.038)       0:00:12.081 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova : Copying over config.json files for nova-api-bootstrap] ************\nSaturday 21 March 2026  01:03:10 +0000 (0:00:00.673)       0:00:12.754 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova : Copying over nova.conf for nova-api-bootstrap] ********************\nSaturday 21 March 2026  01:03:11 +0000 (0:00:01.338)       0:00:14.093 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova : include_tasks] ****************************************************\nSaturday 21 March 2026  01:03:14 +0000 (0:00:02.956)       0:00:17.049 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Running Nova API bootstrap container] *****************************\nSaturday 21 March 2026  01:03:15 +0000 (0:00:00.412)       0:00:17.462 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova : Create cell0 mappings] ********************************************\nSaturday 21 March 2026  01:03:44 +0000 (0:00:29.554)       0:00:47.016 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Get a list of existing cells] ********************************\nSaturday 21 March 2026  01:03:58 +0000 (0:00:14.206)       0:01:01.222 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Extract current cell settings from list] *********************\nSaturday 21 March 2026  01:04:09 +0000 (0:00:10.978)       0:01:12.201 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova : Update cell0 mappings] ********************************************\nSaturday 21 March 2026  01:04:10 +0000 (0:00:00.786)       0:01:12.988 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [nova : include_tasks] ****************************************************\nSaturday 21 March 2026  01:04:10 +0000 (0:00:00.412)       0:01:13.400 ******** \n\u001b[0;36mincluded: /ansible/roles/nova/tasks/bootstrap_service.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova : Running Nova API bootstrap container] *****************************\nSaturday 21 March 2026  01:04:11 +0000 (0:00:00.721)       0:01:14.121 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Bootstrap upgrade] *******************************************************\nSaturday 21 March 2026  01:04:29 +0000 (0:00:18.095)       0:01:32.216 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Bootstrap nova cell databases] *******************************************\n\nTASK [Bootstrap deploy] ********************************************************\nSaturday 21 March 2026  01:04:30 +0000 (0:00:00.288)       0:01:32.505 ******** \n\u001b[0;36mincluded: nova-cell for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova-cell : Creating Nova cell database] *********************************\nSaturday 21 March 2026  01:04:30 +0000 (0:00:00.629)       0:01:33.135 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Creating Nova cell database user and setting permissions] ****\nSaturday 21 March 2026  01:04:32 +0000 (0:00:01.747)       0:01:34.883 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [service-rabbitmq : nova | Ensure RabbitMQ vhosts exist] ******************\nSaturday 21 March 2026  01:04:34 +0000 (0:00:01.739)       0:01:36.622 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-rabbitmq : nova | Ensure RabbitMQ users exist] *******************\nSaturday 21 March 2026  01:04:34 +0000 (0:00:00.407)       0:01:37.029 ******** \n\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> {{ service_rabbitmq_delegate_host }}]\u001b[0m\n\nTASK [service-rabbitmq : nova | Ensure RabbitMQ vhosts exist] ******************\nSaturday 21 March 2026  01:04:40 +0000 (0:00:06.230)       0:01:43.259 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-rabbitmq : nova | Ensure RabbitMQ users exist] *******************\nSaturday 21 March 2026  01:04:41 +0000 (0:00:00.507)       0:01:43.767 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Ensuring config directories exist] ***************************\nSaturday 21 March 2026  01:04:42 +0000 (0:00:01.435)       0:01:45.203 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Copying over config.json files for nova-cell-bootstrap] ******\nSaturday 21 March 2026  01:04:43 +0000 (0:00:00.479)       0:01:45.682 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Copying over nova.conf for nova-cell-bootstrap] **************\nSaturday 21 March 2026  01:04:44 +0000 (0:00:00.921)       0:01:46.604 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Running Nova cell bootstrap container] ***********************\nSaturday 21 March 2026  01:04:46 +0000 (0:00:02.000)       0:01:48.604 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Get a list of existing cells] ********************************\nSaturday 21 March 2026  01:05:06 +0000 (0:00:20.161)       0:02:08.766 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Extract current cell settings from list] *********************\nSaturday 21 March 2026  01:05:18 +0000 (0:00:11.738)       0:02:20.504 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Create cell] *************************************************\nSaturday 21 March 2026  01:05:19 +0000 (0:00:01.029)       0:02:21.534 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Update cell] *************************************************\nSaturday 21 March 2026  01:05:32 +0000 (0:00:13.315)       0:02:34.850 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Bootstrap upgrade] *******************************************************\nSaturday 21 March 2026  01:05:33 +0000 (0:00:01.107)       0:02:35.957 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role nova] *********************************************************\n\nTASK [nova : include_tasks] ****************************************************\nSaturday 21 March 2026  01:05:33 +0000 (0:00:00.265)       0:02:36.223 ******** \n\u001b[0;36mincluded: /ansible/roles/nova/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : nova | Creating services] **************************\nSaturday 21 March 2026  01:05:34 +0000 (0:00:00.562)       0:02:36.786 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy (compute_legacy)) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova (compute))\u001b[0m\n\nTASK [service-ks-register : nova | Creating endpoints] *************************\nSaturday 21 March 2026  01:05:37 +0000 (0:00:03.189)       0:02:39.976 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy -> https://api-int.testbed.osism.xyz:8774/v2/%(tenant_id)s -> internal) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy -> https://api.testbed.osism.xyz:8774/v2/%(tenant_id)s -> public) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> https://api-int.testbed.osism.xyz:8774/v2.1 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> https://api.testbed.osism.xyz:8774/v2.1 -> public)\u001b[0m\n\nTASK [service-ks-register : nova | Creating projects] **************************\nSaturday 21 March 2026  01:05:43 +0000 (0:00:06.005)       0:02:45.981 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : nova | Creating users] *****************************\nSaturday 21 March 2026  01:05:46 +0000 (0:00:02.600)       0:02:48.581 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : nova | Creating roles] *****************************\nSaturday 21 March 2026  01:05:49 +0000 (0:00:03.382)       0:02:51.964 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : nova | Granting user roles] ************************\nSaturday 21 March 2026  01:05:52 +0000 (0:00:02.783)       0:02:54.748 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service -> admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service -> service)\u001b[0m\n\nTASK [nova : Ensuring config directories exist] ********************************\nSaturday 21 March 2026  01:05:58 +0000 (0:00:05.976)       0:03:00.725 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova : Check if policies shall be overwritten] ***************************\nSaturday 21 March 2026  01:06:00 +0000 (0:00:01.761)       0:03:02.487 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [nova : Set nova policy file] *********************************************\nSaturday 21 March 2026  01:06:00 +0000 (0:00:00.115)       0:03:02.603 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Check for vendordata file] ****************************************\nSaturday 21 March 2026  01:06:00 +0000 (0:00:00.311)       0:03:02.914 ******** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [nova : Set vendordata file path] *****************************************\nSaturday 21 March 2026  01:06:01 +0000 (0:00:00.737)       0:03:03.651 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : include_tasks] ****************************************************\nSaturday 21 March 2026  01:06:01 +0000 (0:00:00.286)       0:03:03.937 ******** \n\u001b[0;36mincluded: /ansible/roles/nova/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over extra CA certificates] ***********\nSaturday 21 March 2026  01:06:02 +0000 (0:00:00.730)       0:03:04.668 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over backend internal TLS certificate] ***\nSaturday 21 March 2026  01:06:04 +0000 (0:00:02.001)       0:03:06.669 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over backend internal TLS key] ********\nSaturday 21 March 2026  01:06:04 +0000 (0:00:00.570)       0:03:07.240 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Copying over config.json files for services] **********************\nSaturday 21 March 2026  01:06:05 +0000 (0:00:00.923)       0:03:08.164 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova : Copying over nova.conf] *******************************************\nSaturday 21 March 2026  01:06:07 +0000 (0:00:02.111)       0:03:10.276 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova : Copying over existing policy file] ********************************\nSaturday 21 March 2026  01:06:13 +0000 (0:00:05.140)       0:03:15.416 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Copying over nova-api-wsgi.conf] **********************************\nSaturday 21 March 2026  01:06:13 +0000 (0:00:00.591)       0:03:16.008 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [nova : Copying over vendordata file] *************************************\nSaturday 21 March 2026  01:06:15 +0000 (0:00:01.643)       0:03:17.652 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Check nova containers] ********************************************\nSaturday 21 March 2026  01:06:15 +0000 (0:00:00.306)       0:03:17.959 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova : Flush handlers] ***************************************************\nSaturday 21 March 2026  01:06:17 +0000 (0:00:01.688)       0:03:19.647 ******** \n\nTASK [nova : Flush handlers] ***************************************************\nSaturday 21 March 2026  01:06:17 +0000 (0:00:00.144)       0:03:19.792 ******** \n\nTASK [nova : Flush handlers] ***************************************************\nSaturday 21 March 2026  01:06:17 +0000 (0:00:00.131)       0:03:19.923 ******** \n\nRUNNING HANDLER [nova : Restart nova-scheduler container] **********************\nSaturday 21 March 2026  01:06:17 +0000 (0:00:00.302)       0:03:20.226 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [nova : Restart nova-api container] ****************************\nSaturday 21 March 2026  01:06:31 +0000 (0:00:13.745)       0:03:33.971 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY [Apply role nova-cell] ****************************************************\n\nTASK [nova-cell : include_tasks] ***********************************************\nSaturday 21 March 2026  01:06:37 +0000 (0:00:05.777)       0:03:39.749 ******** \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova-cell : include_tasks] ***********************************************\nSaturday 21 March 2026  01:06:38 +0000 (0:00:01.178)       0:03:40.927 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Load and persist br_netfilter module] ************************************\nSaturday 21 March 2026  01:06:39 +0000 (0:00:00.706)       0:03:41.634 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: module-load for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [module-load : Load modules] **********************************************\nSaturday 21 March 2026  01:06:40 +0000 (0:00:00.800)       0:03:42.434 ******** \n\u001b[0;32mok: [testbed-node-3] => (item=br_netfilter)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=br_netfilter)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=br_netfilter)\u001b[0m\n\nTASK [module-load : Persist modules via modules-load.d] ************************\nSaturday 21 March 2026  01:06:41 +0000 (0:00:01.205)       0:03:43.639 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=br_netfilter)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=br_netfilter)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=br_netfilter)\u001b[0m\n\nTASK [module-load : Drop module persistence] ***********************************\nSaturday 21 March 2026  01:06:42 +0000 (0:00:01.110)       0:03:44.750 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=br_netfilter) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=br_netfilter) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=br_netfilter) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Enable bridge-nf-call sysctl variables] **********************\nSaturday 21 March 2026  01:06:42 +0000 (0:00:00.519)       0:03:45.270 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m\n\nTASK [nova-cell : Install udev kolla kvm rules] ********************************\nSaturday 21 March 2026  01:06:44 +0000 (0:00:02.097)       0:03:47.367 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Mask qemu-kvm service] ***************************************\nSaturday 21 March 2026  01:06:45 +0000 (0:00:01.006)       0:03:48.373 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [nova-cell : Ensuring config directories exist] ***************************\nSaturday 21 March 2026  01:06:47 +0000 (0:00:01.601)       0:03:49.975 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova-cell : include_tasks] ***********************************************\nSaturday 21 March 2026  01:06:49 +0000 (0:00:02.252)       0:03:52.228 ******** \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/copy-certs.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over extra CA certificates] ***********\nSaturday 21 March 2026  01:06:50 +0000 (0:00:01.013)       0:03:53.241 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over backend internal TLS certificate] ***\nSaturday 21 March 2026  01:06:53 +0000 (0:00:03.140)       0:03:56.382 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over backend internal TLS key] ********\nSaturday 21 March 2026  01:06:55 +0000 (0:00:01.439)       0:03:57.821 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvir\u001b[32m2026-03-21 01:10:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\ntd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : include_tasks] ***********************************************\nSaturday 21 March 2026  01:06:57 +0000 (0:00:02.005)       0:03:59.827 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/external_ceph.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [nova-cell : Check nova keyring file] *************************************\nSaturday 21 March 2026  01:06:58 +0000 (0:00:00.979)       0:04:00.807 ******** \n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [nova-cell : Check cinder keyring file] ***********************************\nSaturday 21 March 2026  01:06:59 +0000 (0:00:00.847)       0:04:01.654 ******** \n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [nova-cell : Extract nova key from file] **********************************\nSaturday 21 March 2026  01:07:00 +0000 (0:00:00.951)       0:04:02.606 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Extract cinder key from file] ********************************\nSaturday 21 March 2026  01:07:00 +0000 (0:00:00.509)       0:04:03.115 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Copy over ceph nova keyring file] ****************************\nSaturday 21 March 2026  01:07:01 +0000 (0:00:00.458)       0:04:03.574 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m\n\nTASK [nova-cell : Copy over ceph cinder keyring file] **************************\nSaturday 21 March 2026  01:07:02 +0000 (0:00:00.981)       0:04:04.555 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m\n\nTASK [nova-cell : Copy over ceph.conf] *****************************************\nSaturday 21 March 2026  01:07:03 +0000 (0:00:01.137)       0:04:05.693 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=nova-libvirt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=nova-libvirt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=nova-libvirt)\u001b[0m\n\nTASK [nova-cell : Ensure /etc/ceph directory exists (host libvirt)] ************\nSaturday 21 March 2026  01:07:06 +0000 (0:00:03.282)       0:04:08.975 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Copy over ceph.conf (host libvirt)] **************************\nSaturday 21 March 2026  01:07:06 +0000 (0:00:00.288)       0:04:09.264 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Ensuring libvirt secrets directory exists] *******************\nSaturday 21 March 2026  01:07:07 +0000 (0:00:00.284)       0:04:09.549 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Pushing nova secret xml for libvirt] *************************\nSaturday 21 March 2026  01:07:08 +0000 (0:00:01.433)       0:04:10.982 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'})\u001b[0m\n\nTASK [nova-cell : Pushing secrets key for libvirt] *****************************\nSaturday 21 March 2026  01:07:11 +0000 (0:00:03.333)       0:04:14.316 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Include tasks from qemu_wrapper.yml] *************************\nSaturday 21 March 2026  01:07:14 +0000 (0:00:03.032)       0:04:17.349 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/qemu_wrapper.yml for testbed-node-4, testbed-node-3, testbed-node-5\u001b[0m\n\nTASK [nova-cell : Check qemu wrapper file] *************************************\nSaturday 21 March 2026  01:07:16 +0000 (0:00:01.743)       0:04:19.093 ******** \n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [nova-cell : Copy qemu wrapper] *******************************************\nSaturday 21 March 2026  01:07:17 +0000 (0:00:01.070)       0:04:20.163 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Check if policies shall be overwritten] **********************\nSaturday 21 March 2026  01:07:18 +0000 (0:00:00.323)       0:04:20.487 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [nova-cell : Set nova policy file] ****************************************\nSaturday 21 March 2026  01:07:18 +0000 (0:00:00.124)       0:04:20.611 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Check for vendordata file] ***********************************\nSaturday 21 March 2026  01:07:18 +0000 (0:00:00.772)       0:04:21.384 ******** \n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\nTASK [nova-cell : Set vendordata file path] ************************************\nSaturday 21 March 2026  01:07:19 +0000 (0:00:00.737)       0:04:22.121 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Copying over config.json files for services] *****************\nSaturday 21 March 2026  01:07:20 +0000 (0:00:00.556)       0:04:22.678 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova-cell : Copying over nova.conf] **************************************\nSaturday 21 March 2026  01:07:23 +0000 (0:00:03.684)       0:04:26.362 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova-cell : Copying over Nova compute provider config] *******************\nSaturday 21 March 2026  01:07:29 +0000 (0:00:05.753)       0:04:32.116 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Copying over libvirt configuration] **************************\nSaturday 21 March 2026  01:07:31 +0000 (0:00:01.650)       0:04:33.766 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m\n\nTASK [nova-cell : Copying over libvirt TLS keys] *******************************\nSaturday 21 March 2026  01:07:34 +0000 (0:00:03.479)       0:04:37.246 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Copying over libvirt SASL configuration] *********************\nSaturday 21 March 2026  01:07:35 +0000 (0:00:00.732)       0:04:37.979 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m\n\nTASK [nova-cell : Copying files for nova-ssh] **********************************\nSaturday 21 March 2026  01:07:40 +0000 (0:00:05.150)       0:04:43.129 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\nTASK [nova-cell : Copying VMware vCenter CA file] ******************************\nSaturday 21 March 2026  01:07:46 +0000 (0:00:06.221)       0:04:49.351 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Copying 'release' file for nova_compute] *********************\nSaturday 21 March 2026  01:07:47 +0000 (0:00:00.557)       0:04:49.908 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Generating 'hostnqn' file for nova_compute] ******************\nSaturday 21 March 2026  01:07:48 +0000 (0:00:00.796)       0:04:50.705 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Generating 'hostid' file for nova_compute] *******************\nSaturday 21 March 2026  01:07:50 +0000 (0:00:01.858)       0:04:52.563 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [nova-cell : Copying over existing policy file] ***************************\nSaturday 21 March 2026  01:07:52 +0000 (0:00:02.090)       0:04:54.654 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Copying over vendordata file to containers] ******************\nSaturday 21 March 2026  01:07:53 +0000 (0:00:01.328)       0:04:55.983 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Check nova-cell containers] **********************************\nSaturday 21 March 2026  01:07:54 +0000 (0:00:00.784)       0:04:56.767 ******** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova-cell : include_tasks] ***********************************************\nSaturday 21 March 2026  01:07:56 +0000 (0:00:02.537)       0:04:59.305 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Flush handlers] **********************************************\nSaturday 21 March 2026  01:07:57 +0000 (0:00:00.761)       0:05:00.067 ******** \n\nTASK [nova-cell : Flush handlers] **********************************************\nSaturday 21 March 2026  01:07:57 +0000 (0:00:00.128)       0:05:00.195 ******** \n\nTASK [nova-cell : Flush handlers] **********************************************\nSaturday 21 March 2026  01:07:57 +0000 (0:00:00.127)       0:05:00.322 ******** \n\nTASK [nova-cell : Flush handlers] **********************************************\nSaturday 21 March 2026  01:07:58 +0000 (0:00:00.143)       0:05:00.466 ******** \n\nTASK [nova-cell : Flush handlers] **********************************************\nSaturday 21 March 2026  01:07:58 +0000 (0:00:00.139)       0:05:00.605 ******** \n\nTASK [nova-cell : Flush handlers] **********************************************\nSaturday 21 March 2026  01:07:58 +0000 (0:00:00.277)       0:05:00.883 ******** \n\nRUNNING HANDLER [nova-cell : Restart nova-conductor container] *****************\nSaturday 21 March 2026  01:07:58 +0000 (0:00:00.144)       0:05:01.028 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Restart nova-novncproxy container] ****************\nSaturday 21 March 2026  01:08:05 +0000 (0:00:06.765)       0:05:07.794 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Restart nova-ssh container] ***********************\nSaturday 21 March 2026  01:08:21 +0000 (0:00:16.178)       0:05:23.973 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Restart nova-libvirt container] *******************\nSaturday 21 March 2026  01:08:40 +0000 (0:00:19.195)       0:05:43.169 ******** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Checking libvirt container is ready] **************\nSaturday 21 March 2026  01:09:05 +0000 (0:00:24.643)       0:06:07.812 ******** \n\u001b[1;30mFAILED - RETRYING: [testbed-node-3]: Checking libvirt container is ready (10 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-4]: Checking libvirt container is ready (10 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-5]: Checking libvirt container is ready (10 retries left).\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Create libvirt SASL user] *************************\nSaturday 21 March 2026  01:09:11 +0000 (0:00:06.091)       0:06:13.904 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Restart nova-compute container] *******************\nSaturday 21 March 2026  01:09:12 +0000 (0:00:00.718)       0:06:14.622 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Wait for nova-compute services to update service versions] ***\nSaturday 21 March 2026  01:09:34 +0000 (0:00:21.795)       0:06:36.418 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [nova-cell : Waiting for nova-compute services to register themselves] ****\nSaturday 21 March 2026  01:09:34 +0000 (0:00:00.237)       0:06:36.655 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-3 -> testbed-node-0]: Waiting for nova-compute services to register themselves (20 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [nova-cell : Fail if nova-compute service failed to register] *************\nSaturday 21 March 2026  01:09:54 +0000 (0:00:20.641)       0:06:57.296 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [nova-cell : Include discover_computes.yml] *******************************\nSaturday 21 March 2026  01:10:03 +0000 (0:00:08.171)       0:07:05.468 ******** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/discover_computes.yml for testbed-node-3\u001b[0m\n\nTASK [nova-cell : Get a list of existing cells] ********************************\nSaturday 21 March 2026  01:10:06 +0000 (0:00:03.781)       0:07:09.249 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [nova-cell : Extract current cell settings from list] *********************\nSaturday 21 March 2026  01:10:20 +0000 (0:00:14.062)       0:07:23.312 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [nova-cell : Fail if cell settings not found] *****************************\nSaturday 21 March 2026  01:10:22 +0000 (0:00:01.490)       0:07:24.803 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [nova-cell : Discover nova hosts] *****************************************\nSaturday 21 March 2026  01:10:23 +0000 (0:00:01.339)       0:07:26.142 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [nova-cell : Remove old nova_libvirt_secrets container volume] ************\nSaturday 21 March 2026  01:10:36 +0000 (0:00:12.809)       0:07:38.952 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Refresh nova scheduler cell cache] ***************************************\n\nTASK [nova : Refresh cell cache in nova scheduler] *****************************\nSaturday 21 March 2026  01:10:38 +0000 (0:00:01.930)       0:07:40.883 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Reload global Nova super conductor services] *****************************\n\nTASK [nova : Reload nova super conductor services to remove RPC version pin] ***\nSaturday 21 March 2026  01:10:39 +0000 (0:00:01.021)       0:07:41.904 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Reload Nova cell services] ***********************************************\n\nTASK [nova-cell : Reload nova cell services to remove RPC version cap] *********\nSaturday 21 March 2026  01:10:39 +0000 (0:00:00.485)       0:07:42.389 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Reload global Nova API services] *****************************************\n\nTASK [nova : Reload nova API services to remove RPC version pin] ***************\nSaturday 21 March 2026  01:10:41 +0000 (0:00:01.266)       0:07:43.656 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-scheduler) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-api) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-scheduler) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-api) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-scheduler) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-api) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Run Nova API online data migrations] *************************************\n\nTASK [nova : Run Nova API online database migrations] **************************\nSaturday 21 March 2026  01:10:41 +0000 (0:00:00.668)       0:07:44.324 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY [Run Nova cell online data migrations] ************************************\n\nTASK [nova-cell : Run Nova cell online database migrations] ********************\nSaturday 21 March 2026  01:10:42 +0000 (0:00:00.629)       0:07:44.954 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=54  \u001b[0m \u001b[0;33mchanged=35  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=46  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=53  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=53  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=46  \u001b[0m \u001b[0;33mchanged=28  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=21  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=40  \u001b[0m \u001b[0;33mchanged=28  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=40  \u001b[0m \u001b[0;33mchanged=28  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  01:10:42 +0000 (0:00:00.411)       0:07:45.366 ******** \n=============================================================================== \nnova : Running Nova API bootstrap container ---------------------------- 29.55s\nnova-cell : Restart nova-libvirt container ----------------------------- 24.64s\nnova-cell : Restart nova-compute container ----------------------------- 21.80s\nnova-cell : Waiting for nova-compute services to register themselves --- 20.64s\nnova-cell : Running Nova cell bootstrap container ---------------------- 20.16s\nnova-cell : Restart nova-ssh container --------------------------------- 19.20s\nnova : Running Nova API bootstrap container ---------------------------- 18.10s\nnova-cell : Restart nova-novncproxy container -------------------------- 16.18s\nnova : Create cell0 mappings ------------------------------------------- 14.21s\nnova-cell : Get a list of existing cells ------------------------------- 14.06s\nnova : Restart nova-scheduler container -------------------------------- 13.75s\nnova-cell : Create cell ------------------------------------------------ 13.32s\nnova-cell : Discover nova hosts ---------------------------------------- 12.81s\nnova-cell : Get a list of existing cells ------------------------------- 11.74s\nnova-cell : Get a list of existing cells ------------------------------- 10.98s\nnova-cell : Fail if nova-compute service failed to register ------------- 8.17s\nnova-cell : Restart nova-conductor container ---------------------------- 6.77s\nservice-rabbitmq : nova | Ensure RabbitMQ users exist ------------------- 6.23s\nnova-cell : Copying files for nova-ssh ---------------------------------- 6.22s\nnova-cell : Checking libvirt container is ready ------------------------- 6.09s\n\u001b[32m2026-03-21 01:10:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:10:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:10:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:10:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:10:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:11:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:11:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:11:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:11:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:11:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:11:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:11:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:11:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:11:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m\n\u001b[32m2026-03-21 01:11:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-21 01:11:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  01:06:43 +0000 (0:00:00.312)       0:00:00.312 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  01:06:43 +0000 (0:00:00.281)       0:00:00.594 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_octavia_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_octavia_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_octavia_True)\u001b[0m\n\nPLAY [Apply role octavia] ******************************************************\n\nTASK [octavia : include_tasks] *************************************************\nSaturday 21 March 2026  01:06:43 +0000 (0:00:00.294)       0:00:00.888 ******** \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : octavia | Creating services] ***********************\nSaturday 21 March 2026  01:06:44 +0000 (0:00:00.622)       0:00:01.510 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia (load-balancer))\u001b[0m\n\nTASK [service-ks-register : octavia | Creating endpoints] **********************\nSaturday 21 March 2026  01:06:47 +0000 (0:00:03.691)       0:00:05.202 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> https://api-int.testbed.osism.xyz:9876 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> https://api.testbed.osism.xyz:9876 -> public)\u001b[0m\n\nTASK [service-ks-register : octavia | Creating projects] ***********************\nSaturday 21 March 2026  01:06:54 +0000 (0:00:06.475)       0:00:11.678 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : octavia | Creating users] **************************\nSaturday 21 March 2026  01:06:57 +0000 (0:00:02.795)       0:00:14.473 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : octavia | Creating roles] **************************\nSaturday 21 March 2026  01:07:04 +0000 (0:00:07.483)       0:00:21.957 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : octavia | Granting user roles] *********************\nSaturday 21 March 2026  01:07:08 +0000 (0:00:03.400)       0:00:25.357 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service -> admin)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=octavia -> service -> admin)\u001b[0m\n\nTASK [octavia : Adding octavia related roles] **********************************\nSaturday 21 March 2026  01:07:16 +0000 (0:00:07.991)       0:00:33.348 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_observer)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_global_observer)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_member)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_quota_admin)\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nSaturday 21 March 2026  01:07:30 +0000 (0:00:14.921)       0:00:48.270 ******** \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/prepare.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [octavia : Create amphora flavor] *****************************************\nSaturday 21 March 2026  01:07:31 +0000 (0:00:00.836)       0:00:49.107 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Create nova keypair for amphora] *******************************\nSaturday 21 March 2026  01:07:36 +0000 (0:00:04.295)       0:00:53.402 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Get service project id] ****************************************\nSaturday 21 March 2026  01:07:40 +0000 (0:00:04.471)       0:00:57.874 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Create security groups for octavia] ****************************\nSaturday 21 March 2026  01:07:43 +0000 (0:00:02.994)       0:01:00.869 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=lb-mgmt-sec-grp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=lb-health-mgr-sec-grp)\u001b[0m\n\nTASK [octavia : Add rules for security groups] *********************************\nSaturday 21 March 2026  01:07:52 +0000 (0:00:08.546)       0:01:09.415 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'icmp'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'tcp', 'src_port': 22, 'dst_port': 22}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'tcp', 'src_port': '9443', 'dst_port': '9443'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-health-mgr-sec-grp', 'enabled': True}, {'protocol': 'udp', 'src_port': '5555', 'dst_port': '5555'}])\u001b[0m\n\nTASK [octavia : Create loadbalancer management network] ************************\nSaturday 21 March 2026  01:08:07 +0000 (0:00:15.225)       0:01:24.641 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Create loadbalancer management subnet] *************************\nSaturday 21 March 2026  01:08:11 +0000 (0:00:03.946)       0:01:28.587 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Create loadbalancer management router for IPv6] ****************\nSaturday 21 March 2026  01:08:15 +0000 (0:00:04.377)       0:01:32.965 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Update loadbalancer management subnet] *************************\nSaturday 21 March 2026  01:08:15 +0000 (0:00:00.219)       0:01:33.185 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nSaturday 21 March 2026  01:08:20 +0000 (0:00:04.931)       0:01:38.117 ******** \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/hm-interface.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [octavia : Create ports for Octavia health-manager nodes] *****************\nSaturday 21 March 2026  01:08:21 +0000 (0:00:00.796)       0:01:38.913 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Update Octavia health manager port host_id] ********************\nSaturday 21 March 2026  01:08:27 +0000 (0:00:05.439)       0:01:44.353 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [octavia : Add Octavia port to openvswitch br-int] ************************\nSaturday 21 March 2026  01:08:31 +0000 (0:00:04.231)       0:01:48.584 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Install isc-dhcp-client package] *******************************\nSaturday 21 March 2026  01:08:32 +0000 (0:00:00.723)       0:01:49.308 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Create octavia dhclient conf] **********************************\nSaturday 21 March 2026  01:08:33 +0000 (0:00:01.411)       0:01:50.719 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Create octavia-interface service] ******************************\nSaturday 21 March 2026  01:08:34 +0000 (0:00:01.141)       0:01:51.860 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Restart octavia-interface.service if required] *****************\nSaturday 21 March 2026  01:08:35 +0000 (0:00:01.112)       0:01:52.973 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Enable and start octavia-interface.service] ********************\nSaturday 21 March 2026  01:08:37 +0000 (0:00:02.052)       0:01:55.026 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Wait for interface ohm0 ip appear] *****************************\nSaturday 21 March 2026  01:08:39 +0000 (0:00:01.436)       0:01:56.462 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Gather facts] **************************************************\nSaturday 21 March 2026  01:08:39 +0000 (0:00:00.530)       0:01:56.992 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nSaturday 21 March 2026  01:08:41 +0000 (0:00:02.275)       0:01:59.268 ******** \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/get_resources_info.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [octavia : Get amphora flavor info] ***************************************\nSaturday 21 March 2026  01:08:42 +0000 (0:00:00.688)       0:01:59.957 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Get service project id] ****************************************\nSaturday 21 March 2026  01:08:46 +0000 (0:00:03.444)       0:02:03.402 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Get security groups for octavia] *******************************\nSaturday 21 March 2026  01:08:49 +0000 (0:00:02.931)       0:02:06.334 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=lb-mgmt-sec-grp)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=lb-health-mgr-sec-grp)\u001b[0m\n\nTASK [octavia : Get loadbalancer management network] ***************************\nSaturday 21 March 2026  01:08:56 +0000 (0:00:07.195)       0:02:13.530 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Set octavia resources facts] ***********************************\nSaturday 21 March 2026  01:08:59 +0000 (0:00:03.095)       0:02:16.625 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Ensuring config directories exist] *****************************\nSaturday 21 March 2026  01:08:59 +0000 (0:00:00.333)       0:02:16.958 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [octavia : Check if policies shall be overwritten] ************************\nSaturday 21 March 2026  01:09:02 +0000 (0:00:02.475)       0:02:19.434 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Set octavia policy file] ***************************************\nSaturday 21 March 2026  01:09:02 +0000 (0:00:00.128)       0:02:19.563 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Copying over existing policy file] *****************************\nSaturday 21 March 2026  01:09:02 +0000 (0:00:00.268)       0:02:19.831 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nSaturday 21 March 2026  01:09:03 +0000 (0:00:00.663)       0:02:20.494 ******** \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : octavia | Copying over extra CA certificates] ********\nSaturday 21 March 2026  01:09:03 +0000 (0:00:00.666)       0:02:21.161 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : octavia | Copying over backend internal TLS certificate] ***\nSaturday 21 March 2026  01:09:08 +0000 (0:00:04.239)       0:02:25.400 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : octavia | Copying over backend internal TLS key] *****\nSaturday 21 March 2026  01:09:08 +0000 (0:00:00.678)       0:02:26.079 ******** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Copying over config.json files for services] *******************\nSaturday 21 March 2026  01:09:09 +0000 (0:00:01.021)       0:02:27.101 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [octavia : Copying over octavia-wsgi.conf] ********************************\nSaturday 21 March 2026  01:09:14 +0000 (0:00:05.045)       0:02:32.146 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m\n\nTASK [octavia : Copying over octavia.conf] *************************************\nSaturday 21 March 2026  01:09:16 +0000 (0:00:01.596)       0:02:33.742 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [octavia : Copying over Octavia SSH key] **********************************\nSaturday 21 March 2026  01:09:32 +0000 (0:00:16.466)       0:02:50.209 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Copying certificate files for octavia-worker] ******************\nSaturday 21 March 2026  01:09:34 +0000 (0:00:02.042)       0:02:52.251 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m\n\nTASK [octavia : Copying certificate files for octavia-housekeeping] ************\nSaturday 21 March 2026  01:09:40 +0000 (0:00:05.950)       0:02:58.202 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m\n\nTASK [octavia : Copying certificate files for octavia-health-manager] **********\nSaturday 21 March 2026  01:09:45 +0000 (0:00:05.026)       0:03:03.229 ******** \n\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m\n\nTASK [octavia : Check octavia containers] **************************************\nSaturday 21 March 2026  01:09:51 +0000 (0:00:05.076)       0:03:08.306 ******** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nSaturday 21 March 2026  01:09:54 +0000 (0:00:03.299)       0:03:11.605 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Creating Octavia database] *************************************\nSaturday 21 March 2026  01:09:54 +0000 (0:00:00.455)       0:03:12.061 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Creating Octavia persistence database] *************************\nSaturday 21 March 2026  01:09:56 +0000 (0:00:01.896)       0:03:13.957 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Creating Octavia database user and setting permissions] ********\nSaturday 21 March 2026  01:09:58 +0000 (0:00:01.981)       0:03:15.938 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Creating Octavia persistence database user and setting permissions] ***\nSaturday 21 March 2026  01:10:00 +0000 (0:00:02.202)       0:03:18.141 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Running Octavia bootstrap container] ***************************\nSaturday 21 March 2026  01:10:03 +0000 (0:00:02.440)       0:03:20.582 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Flush handlers] ************************************************\nSaturday 21 March 2026  01:10:24 +0000 (0:00:21.649)       0:03:42.231 ******** \n\nTASK [octavia : Flush handlers] ************************************************\nSaturday 21 March 2026  01:10:24 +0000 (0:00:00.069)       0:03:42.301 ******** \n\nTASK [octavia : Flush handlers] ************************************************\nSaturday 21 March 2026  01:10:25 +0000 (0:00:00.066)       0:03:42.368 ******** \n\nRUNNING HANDLER [octavia : Restart octavia-api container] **********************\nSaturday 21 March 2026  01:10:25 +0000 (0:00:00.066)       0:03:42.434 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [octavia : Restart octavia-driver-agent container] *************\nSaturday 21 March 2026  01:10:35 +0000 (0:00:10.446)       0:03:52.880 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [octavia : Restart octavia-health-manager container] ***********\nSaturday 21 March 2026  01:10:45 +0000 (0:00:10.342)       0:04:03.223 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [octavia : Restart octavia-housekeeping container] *************\nSaturday 21 March 2026  01:10:55 +0000 (0:00:10.042)       0:04:13.266 ******** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [octavia : Restart octavia-worker container] *******************\nSaturday 21 March 2026  01:11:04 +0000 (0:00:08.205)       0:04:21.471 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=57  \u001b[0m \u001b[0;33mchanged=38  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=33  \u001b[0m \u001b[0;33mchanged=22  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=33  \u001b[0m \u001b[0;33mchanged=22  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  01:11:14 +0000 (0:00:10.161)       0:04:31.633 ******** \n=============================================================================== \noctavia : Running Octavia bootstrap container -------------------------- 21.65s\noctavia : Copying over octavia.conf ------------------------------------ 16.47s\noctavia : Add rules for security groups -------------------------------- 15.23s\noctavia : Adding octavia related roles --------------------------------- 14.92s\noctavia : Restart octavia-api container -------------------------------- 10.45s\noctavia : Restart octavia-driver-agent container ----------------------- 10.34s\noctavia : Restart octavia-worker container ----------------------------- 10.16s\noctavia : Restart octavia-health-manager container --------------------- 10.04s\noctavia : Create security groups for octavia ---------------------------- 8.55s\noctavia : Restart octavia-housekeeping container ------------------------ 8.21s\nservice-ks-register : octavia | Granting user roles --------------------- 7.99s\nservice-ks-register : octavia | Creating users -------------------------- 7.48s\noctavia : Get security groups for octavia ------------------------------- 7.20s\nservice-ks-register : octavia | Creating endpoints ---------------------- 6.48s\noctavia : Copying certificate files for octavia-worker ------------------ 5.95s\noctavia : Create ports for Octavia health-manager nodes ----------------- 5.44s\noctavia : Copying certificate files for octavia-health-manager ---------- 5.08s\noctavia : Copying over config.json files for services ------------------- 5.05s\noctavia : Copying certificate files for octavia-housekeeping ------------ 5.03s\noctavia : Update loadbalancer management subnet ------------------------- 4.93s\n\u001b[32m2026-03-21 01:11:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-21 01:11:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-21 01:11:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-21 01:11:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-21 01:11:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-21 01:11:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-21 01:11:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-21 01:11:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-21 01:11:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-21 01:11:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-21 01:11:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-21 01:11:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-21 01:11:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-21 01:11:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-21 01:11:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-21 01:12:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-21 01:12:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-21 01:12:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-21 01:12:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-21 01:12:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\n\n--> DEPLOY IN A NUTSHELL -- END -- Sat Mar 21 01:12:16 UTC 2026",
                            "stdout_lines": [
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "",
                                "# PULL IMAGES",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=true",
                                "++ TEMPEST=true",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ echo",
                                "+ echo '# PULL IMAGES'",
                                "",
                                "+ echo",
                                "++ semver latest 7.0.0",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ osism apply --no-wait -r 2 -e custom pull-images",
                                "\u001b[32m2026-03-21 00:39:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play pull-images in environment custom\u001b[0m",
                                "\u001b[32m2026-03-21 00:39:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of pull-images.\u001b[0m",
                                "\u001b[32m2026-03-21 00:39:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c302b01c-13af-43e8-aa09-d4d9076680ca (pull-images) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 00:39:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c302b01c-13af-43e8-aa09-d4d9076680ca is running in background. No more output. Check ARA for logs.\u001b[0m",
                                "\u001b[32m2026-03-21 00:39:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play wipe-partitions in environment custom\u001b[0m",
                                "\u001b[32m2026-03-21 00:39:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of wipe-partitions.\u001b[0m",
                                "\u001b[32m2026-03-21 00:39:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 619848fa-4eb2-47d1-a281-4cc0601421ae (wipe-partitions) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 00:39:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 619848fa-4eb2-47d1-a281-4cc0601421ae (wipe-partitions) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Wipe partitions] *********************************************************",
                                "",
                                "TASK [Find all logical devices owned by UID 167] *******************************",
                                "Saturday 21 March 2026  00:40:00 +0000 (0:00:00.166)       0:00:00.166 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Remove all rook related logical devices] *********************************",
                                "Saturday 21 March 2026  00:40:01 +0000 (0:00:01.194)       0:00:01.360 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Find all logical devices with prefix ceph] *******************************",
                                "Saturday 21 March 2026  00:40:01 +0000 (0:00:00.218)       0:00:01.579 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Remove all ceph related logical devices] *********************************",
                                "Saturday 21 March 2026  00:40:02 +0000 (0:00:00.531)       0:00:02.110 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Check device availability] ***********************************************",
                                "Saturday 21 March 2026  00:40:02 +0000 (0:00:00.236)       0:00:02.347 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdd)\u001b[0m",
                                "",
                                "TASK [Wipe partitions with wipefs] *********************************************",
                                "Saturday 21 March 2026  00:40:04 +0000 (0:00:01.439)       0:00:03.787 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdd)\u001b[0m",
                                "",
                                "TASK [Overwrite first 32M with zeros] ******************************************",
                                "Saturday 21 March 2026  00:40:05 +0000 (0:00:01.363)       0:00:05.150 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdd)\u001b[0m",
                                "",
                                "TASK [Reload udev rules] *******************************************************",
                                "Saturday 21 March 2026  00:40:08 +0000 (0:00:03.187)       0:00:08.337 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Request device events from the kernel] ***********************************",
                                "Saturday 21 March 2026  00:40:09 +0000 (0:00:00.599)       0:00:08.937 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:40:09 +0000 (0:00:00.601)       0:00:09.538 ******** ",
                                "=============================================================================== ",
                                "Overwrite first 32M with zeros ------------------------------------------ 3.19s",
                                "Check device availability ----------------------------------------------- 1.44s",
                                "Wipe partitions with wipefs --------------------------------------------- 1.36s",
                                "Find all logical devices owned by UID 167 ------------------------------- 1.19s",
                                "Request device events from the kernel ----------------------------------- 0.60s",
                                "Reload udev rules ------------------------------------------------------- 0.60s",
                                "Find all logical devices with prefix ceph ------------------------------- 0.53s",
                                "Remove all ceph related logical devices --------------------------------- 0.24s",
                                "Remove all rook related logical devices --------------------------------- 0.22s",
                                "\u001b[32m2026-03-21 00:40:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of facts.\u001b[0m",
                                "\u001b[32m2026-03-21 00:40:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ed9c7fdc-dbc7-44ff-b976-11256150f150 (facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 00:40:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task ed9c7fdc-dbc7-44ff-b976-11256150f150 (facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role facts] ********************************************************",
                                "",
                                "TASK [osism.commons.facts : Create custom facts directory] *********************",
                                "Saturday 21 March 2026  00:40:25 +0000 (0:00:00.325)       0:00:00.325 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Copy fact files] ***********************************",
                                "Saturday 21 March 2026  00:40:26 +0000 (0:00:01.268)       0:00:01.593 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Saturday 21 March 2026  00:40:27 +0000 (0:00:01.035)       0:00:02.629 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts if using --limit] *****************************",
                                "",
                                "TASK [Gather facts for all hosts] **********************************************",
                                "Saturday 21 March 2026  00:40:32 +0000 (0:00:05.137)       0:00:07.767 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:40:32 +0000 (0:00:00.450)       0:00:08.217 ******** ",
                                "=============================================================================== ",
                                "Gathers facts about hosts ----------------------------------------------- 5.14s",
                                "osism.commons.facts : Create custom facts directory --------------------- 1.27s",
                                "osism.commons.facts : Copy fact files ----------------------------------- 1.04s",
                                "Gather facts for all hosts ---------------------------------------------- 0.45s",
                                "\u001b[32m2026-03-21 00:40:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of ceph-configure-lvm-volumes.\u001b[0m",
                                "\u001b[32m2026-03-21 00:40:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4f593d2c-7953-4545-b013-3439dbcf4afd (ceph-configure-lvm-volumes) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 00:40:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 4f593d2c-7953-4545-b013-3439dbcf4afd (ceph-configure-lvm-volumes) has been started and output is visible here.\u001b[0m",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Ceph configure LVM] ******************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Saturday 21 March 2026  00:40:38 +0000 (0:00:00.258)       0:00:00.258 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Saturday 21 March 2026  00:40:39 +0000 (0:00:00.218)       0:00:00.476 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:39 +0000 (0:00:00.205)       0:00:00.682 ******** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:39 +0000 (0:00:00.317)       0:00:00.999 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:39 +0000 (0:00:00.370)       0:00:01.370 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:40 +0000 (0:00:00.164)       0:00:01.534 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:40 +0000 (0:00:00.172)       0:00:01.707 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:40 +0000 (0:00:00.174)       0:00:01.881 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:40 +0000 (0:00:00.165)       0:00:02.047 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:40 +0000 (0:00:00.197)       0:00:02.244 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:41 +0000 (0:00:00.181)       0:00:02.426 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:41 +0000 (0:00:00.169)       0:00:02.595 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:41 +0000 (0:00:00.392)       0:00:02.988 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_f08311eb-e2ad-4eb7-8da9-ac6d899a2409)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_f08311eb-e2ad-4eb7-8da9-ac6d899a2409)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:41 +0000 (0:00:00.389)       0:00:03.377 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_e03b171b-b829-41a0-a0b6-cc46d85f013d)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_e03b171b-b829-41a0-a0b6-cc46d85f013d)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:42 +0000 (0:00:00.516)       0:00:03.893 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_9fa3c153-0e2a-45ef-8e48-e0095eb937ad)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_9fa3c153-0e2a-45ef-8e48-e0095eb937ad)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:42 +0000 (0:00:00.525)       0:00:04.419 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:43 +0000 (0:00:00.578)       0:00:04.997 ******** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:43 +0000 (0:00:00.329)       0:00:05.327 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:44 +0000 (0:00:00.173)       0:00:05.501 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:44 +0000 (0:00:00.207)       0:00:05.709 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:44 +0000 (0:00:00.187)       0:00:05.896 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:44 +0000 (0:00:00.186)       0:00:06.083 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:44 +0000 (0:00:00.174)       0:00:06.258 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:45 +0000 (0:00:00.175)       0:00:06.434 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:45 +0000 (0:00:00.174)       0:00:06.608 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:45 +0000 (0:00:00.173)       0:00:06.782 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:46 +0000 (0:00:00.797)       0:00:07.579 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:46 +0000 (0:00:00.180)       0:00:07.760 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:46 +0000 (0:00:00.162)       0:00:07.922 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:46 +0000 (0:00:00.194)       0:00:08.116 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set UUIDs for OSD VGs/LVs] ***********************************************",
                                "Saturday 21 March 2026  00:40:46 +0000 (0:00:00.212)       0:00:08.329 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': None})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': None})\u001b[0m",
                                "",
                                "TASK [Generate WAL VG names] ***************************************************",
                                "Saturday 21 March 2026  00:40:47 +0000 (0:00:00.167)       0:00:08.497 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate DB VG names] ****************************************************",
                                "Saturday 21 March 2026  00:40:47 +0000 (0:00:00.138)       0:00:08.635 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate shared DB/WAL VG names] *****************************************",
                                "Saturday 21 March 2026  00:40:47 +0000 (0:00:00.125)       0:00:08.760 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Define lvm_volumes structures] *******************************************",
                                "Saturday 21 March 2026  00:40:47 +0000 (0:00:00.131)       0:00:08.892 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block only)] *****************************",
                                "Saturday 21 March 2026  00:40:47 +0000 (0:00:00.125)       0:00:09.017 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '64dd100a-6bcc-5763-b557-0732c79bfd55'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '5368e13a-d21f-5912-91bd-da537815d9e8'}})\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db)] *****************************",
                                "Saturday 21 March 2026  00:40:47 +0000 (0:00:00.163)       0:00:09.181 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '64dd100a-6bcc-5763-b557-0732c79bfd55'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '5368e13a-d21f-5912-91bd-da537815d9e8'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + wal)] ****************************",
                                "Saturday 21 March 2026  00:40:47 +0000 (0:00:00.148)       0:00:09.330 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '64dd100a-6bcc-5763-b557-0732c79bfd55'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '5368e13a-d21f-5912-91bd-da537815d9e8'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db + wal)] ***********************",
                                "Saturday 21 March 2026  00:40:48 +0000 (0:00:00.337)       0:00:09.667 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '64dd100a-6bcc-5763-b557-0732c79bfd55'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '5368e13a-d21f-5912-91bd-da537815d9e8'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Compile lvm_volumes] *****************************************************",
                                "Saturday 21 March 2026  00:40:48 +0000 (0:00:00.156)       0:00:09.824 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set OSD devices config data] *********************************************",
                                "Saturday 21 March 2026  00:40:48 +0000 (0:00:00.139)       0:00:09.963 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set DB devices config data] **********************************************",
                                "Saturday 21 March 2026  00:40:48 +0000 (0:00:00.145)       0:00:10.108 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set WAL devices config data] *********************************************",
                                "Saturday 21 March 2026  00:40:48 +0000 (0:00:00.119)       0:00:10.228 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set DB+WAL devices config data] ******************************************",
                                "Saturday 21 March 2026  00:40:48 +0000 (0:00:00.130)       0:00:10.358 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print ceph_osd_devices] **************************************************",
                                "Saturday 21 March 2026  00:40:49 +0000 (0:00:00.139)       0:00:10.498 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;32m        \"sdb\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"64dd100a-6bcc-5763-b557-0732c79bfd55\"\u001b[0m",
                                "\u001b[0;32m        },\u001b[0m",
                                "\u001b[0;32m        \"sdc\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"5368e13a-d21f-5912-91bd-da537815d9e8\"\u001b[0m",
                                "\u001b[0;32m        }\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print WAL devices] *******************************************************",
                                "Saturday 21 March 2026  00:40:49 +0000 (0:00:00.141)       0:00:10.640 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print DB devices] ********************************************************",
                                "Saturday 21 March 2026  00:40:49 +0000 (0:00:00.130)       0:00:10.770 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print shared DB/WAL devices] *********************************************",
                                "Saturday 21 March 2026  00:40:49 +0000 (0:00:00.125)       0:00:10.896 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print configuration data] ************************************************",
                                "Saturday 21 March 2026  00:40:49 +0000 (0:00:00.124)       0:00:11.020 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m",
                                "\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;33m            \"sdb\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"64dd100a-6bcc-5763-b557-0732c79bfd55\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            \"sdc\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"5368e13a-d21f-5912-91bd-da537815d9e8\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        },\u001b[0m",
                                "\u001b[0;33m        \"lvm_volumes\": [\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-64dd100a-6bcc-5763-b557-0732c79bfd55\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-5368e13a-d21f-5912-91bd-da537815d9e8\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-5368e13a-d21f-5912-91bd-da537815d9e8\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        ]\u001b[0m",
                                "\u001b[0;33m    }\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "RUNNING HANDLER [Write configuration file] *************************************",
                                "Saturday 21 March 2026  00:40:49 +0000 (0:00:00.207)       0:00:11.228 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "PLAY [Ceph configure LVM] ******************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Saturday 21 March 2026  00:40:52 +0000 (0:00:02.229)       0:00:13.458 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Saturday 21 March 2026  00:40:52 +0000 (0:00:00.249)       0:00:13.708 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:52 +0000 (0:00:00.242)       0:00:13.950 ******** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:52 +0000 (0:00:00.351)       0:00:14.302 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:53 +0000 (0:00:00.197)       0:00:14.500 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:53 +0000 (0:00:00.192)       0:00:14.693 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:53 +0000 (0:00:00.205)       0:00:14.899 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:53 +0000 (0:00:00.189)       0:00:15.088 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:54 +0000 (0:00:00.572)       0:00:15.661 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:54 +0000 (0:00:00.196)       0:00:15.858 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:54 +0000 (0:00:00.206)       0:00:16.064 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:54 +0000 (0:00:00.186)       0:00:16.250 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:55 +0000 (0:00:00.421)       0:00:16.672 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_4e74a544-8b3a-4a05-98a1-b7cc13f13c18)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_4e74a544-8b3a-4a05-98a1-b7cc13f13c18)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:55 +0000 (0:00:00.418)       0:00:17.090 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_666e73fe-87f5-4f65-9ff2-95a038c8b18d)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_666e73fe-87f5-4f65-9ff2-95a038c8b18d)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:56 +0000 (0:00:00.427)       0:00:17.517 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_8ae4aabb-0174-4999-a1d0-4bf78cbb2965)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_8ae4aabb-0174-4999-a1d0-4bf78cbb2965)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:40:56 +0000 (0:00:00.418)       0:00:17.936 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:56 +0000 (0:00:00.307)       0:00:18.244 ******** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:57 +0000 (0:00:00.372)       0:00:18.616 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:57 +0000 (0:00:00.185)       0:00:18.802 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:57 +0000 (0:00:00.560)       0:00:19.363 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:58 +0000 (0:00:00.171)       0:00:19.535 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:58 +0000 (0:00:00.169)       0:00:19.705 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:58 +0000 (0:00:00.230)       0:00:19.935 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:58 +0000 (0:00:00.213)       0:00:20.149 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:58 +0000 (0:00:00.166)       0:00:20.315 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:59 +0000 (0:00:00.193)       0:00:20.508 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:59 +0000 (0:00:00.517)       0:00:21.026 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:59 +0000 (0:00:00.146)       0:00:21.173 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:40:59 +0000 (0:00:00.150)       0:00:21.323 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:41:00 +0000 (0:00:00.143)       0:00:21.466 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set UUIDs for OSD VGs/LVs] ***********************************************",
                                "Saturday 21 March 2026  00:41:00 +0000 (0:00:00.155)       0:00:21.622 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': None})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': None})\u001b[0m",
                                "",
                                "TASK [Generate WAL VG names] ***************************************************",
                                "Saturday 21 March 2026  00:41:00 +0000 (0:00:00.321)       0:00:21.944 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate DB VG names] ****************************************************",
                                "Saturday 21 March 2026  00:41:00 +0000 (0:00:00.094)       0:00:22.038 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate shared DB/WAL VG names] *****************************************",
                                "Saturday 21 March 2026  00:41:00 +0000 (0:00:00.110)       0:00:22.148 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Define lvm_volumes structures] *******************************************",
                                "Saturday 21 March 2026  00:41:00 +0000 (0:00:00.097)       0:00:22.246 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block only)] *****************************",
                                "Saturday 21 March 2026  00:41:00 +0000 (0:00:00.097)       0:00:22.343 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '45b09425-beda-50b4-a83b-1fc50b5adca4'}})\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db)] *****************************",
                                "Saturday 21 March 2026  00:41:01 +0000 (0:00:00.119)       0:00:22.463 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '45b09425-beda-50b4-a83b-1fc50b5adca4'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + wal)] ****************************",
                                "Saturday 21 March 2026  00:41:01 +0000 (0:00:00.112)       0:00:22.576 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '45b09425-beda-50b4-a83b-1fc50b5adca4'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db + wal)] ***********************",
                                "Saturday 21 March 2026  00:41:01 +0000 (0:00:00.111)       0:00:22.687 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '45b09425-beda-50b4-a83b-1fc50b5adca4'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Compile lvm_volumes] *****************************************************",
                                "Saturday 21 March 2026  00:41:01 +0000 (0:00:00.131)       0:00:22.819 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set OSD devices config data] *********************************************",
                                "Saturday 21 March 2026  00:41:01 +0000 (0:00:00.202)       0:00:23.021 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set DB devices config data] **********************************************",
                                "Saturday 21 March 2026  00:41:01 +0000 (0:00:00.184)       0:00:23.206 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set WAL devices config data] *********************************************",
                                "Saturday 21 March 2026  00:41:01 +0000 (0:00:00.101)       0:00:23.308 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set DB+WAL devices config data] ******************************************",
                                "Saturday 21 March 2026  00:41:02 +0000 (0:00:00.223)       0:00:23.532 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print ceph_osd_devices] **************************************************",
                                "Saturday 21 March 2026  00:41:02 +0000 (0:00:00.114)       0:00:23.646 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;32m        \"sdb\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"ed04aa0b-4f5a-523b-a5d2-05072a1b260d\"\u001b[0m",
                                "\u001b[0;32m        },\u001b[0m",
                                "\u001b[0;32m        \"sdc\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"45b09425-beda-50b4-a83b-1fc50b5adca4\"\u001b[0m",
                                "\u001b[0;32m        }\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print WAL devices] *******************************************************",
                                "Saturday 21 March 2026  00:41:02 +0000 (0:00:00.115)       0:00:23.761 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print DB devices] ********************************************************",
                                "Saturday 21 March 2026  00:41:02 +0000 (0:00:00.107)       0:00:23.868 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print shared DB/WAL devices] *********************************************",
                                "Saturday 21 March 2026  00:41:02 +0000 (0:00:00.103)       0:00:23.972 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print configuration data] ************************************************",
                                "Saturday 21 March 2026  00:41:02 +0000 (0:00:00.110)       0:00:24.082 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m",
                                "\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;33m            \"sdb\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"ed04aa0b-4f5a-523b-a5d2-05072a1b260d\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            \"sdc\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"45b09425-beda-50b4-a83b-1fc50b5adca4\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        },\u001b[0m",
                                "\u001b[0;33m        \"lvm_volumes\": [\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-45b09425-beda-50b4-a83b-1fc50b5adca4\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        ]\u001b[0m",
                                "\u001b[0;33m    }\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "RUNNING HANDLER [Write configuration file] *************************************",
                                "Saturday 21 March 2026  00:41:02 +0000 (0:00:00.162)       0:00:24.245 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "PLAY [Ceph configure LVM] ******************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Saturday 21 March 2026  00:41:03 +0000 (0:00:00.916)       0:00:25.161 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Saturday 21 March 2026  00:41:04 +0000 (0:00:00.364)       0:00:25.526 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:41:04 +0000 (0:00:00.494)       0:00:26.020 ******** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:41:04 +0000 (0:00:00.287)       0:00:26.308 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:41:05 +0000 (0:00:00.185)       0:00:26.493 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:41:05 +0000 (0:00:00.174)       0:00:26.668 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:41:05 +0000 (0:00:00.196)       0:00:26.865 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:41:05 +0000 (0:00:00.188)       0:00:27.054 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:41:05 +0000 (0:00:00.161)       0:00:27.215 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:41:05 +0000 (0:00:00.156)       0:00:27.372 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:41:06 +0000 (0:00:00.190)       0:00:27.563 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:41:06 +0000 (0:00:00.187)       0:00:27.750 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:41:06 +0000 (0:00:00.528)       0:00:28.278 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_f76ae8f0-65cc-4a62-b19c-23f8a40fbde6)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_f76ae8f0-65cc-4a62-b19c-23f8a40fbde6)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:41:07 +0000 (0:00:00.672)       0:00:28.951 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_d7bb81ba-6491-4b5a-bb82-f0b49314eac0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_d7bb81ba-6491-4b5a-bb82-f0b49314eac0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:41:07 +0000 (0:00:00.415)       0:00:29.366 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_cdffe5f1-7f0d-460e-a1db-d6846d226144)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_cdffe5f1-7f0d-460e-a1db-d6846d226144)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:41:08 +0000 (0:00:00.371)       0:00:29.738 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:41:08 +0000 (0:00:00.337)       0:00:30.075 ******** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:41:08 +0000 (0:00:00.347)       0:00:30.423 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:41:09 +0000 (0:00:00.210)       0:00:30.634 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:41:09 +0000 (0:00:00.170)       0:00:30.805 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:41:09 +0000 (0:00:00.187)       0:00:30.993 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:41:09 +0000 (0:00:00.182)       0:00:31.175 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:41:09 +0000 (0:00:00.182)       0:00:31.357 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:41:10 +0000 (0:00:00.507)       0:00:31.865 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:41:10 +0000 (0:00:00.231)       0:00:32.097 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:41:10 +0000 (0:00:00.218)       0:00:32.316 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:41:11 +0000 (0:00:00.639)       0:00:32.956 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:41:11 +0000 (0:00:00.167)       0:00:33.123 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:41:11 +0000 (0:00:00.159)       0:00:33.283 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:41:12 +0000 (0:00:00.169)       0:00:33.453 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set UUIDs for OSD VGs/LVs] ***********************************************",
                                "Saturday 21 March 2026  00:41:12 +0000 (0:00:00.167)       0:00:33.620 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': None})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': None})\u001b[0m",
                                "",
                                "TASK [Generate WAL VG names] ***************************************************",
                                "Saturday 21 March 2026  00:41:12 +0000 (0:00:00.150)       0:00:33.770 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate DB VG names] ****************************************************",
                                "Saturday 21 March 2026  00:41:12 +0000 (0:00:00.127)       0:00:33.898 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate shared DB/WAL VG names] *****************************************",
                                "Saturday 21 March 2026  00:41:12 +0000 (0:00:00.116)       0:00:34.014 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Define lvm_volumes structures] *******************************************",
                                "Saturday 21 March 2026  00:41:12 +0000 (0:00:00.100)       0:00:34.115 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block only)] *****************************",
                                "Saturday 21 March 2026  00:41:13 +0000 (0:00:00.317)       0:00:34.432 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'ede169e1-d974-55ed-a938-c303cb89cb5b'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}})\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db)] *****************************",
                                "Saturday 21 March 2026  00:41:13 +0000 (0:00:00.206)       0:00:34.639 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'ede169e1-d974-55ed-a938-c303cb89cb5b'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + wal)] ****************************",
                                "Saturday 21 March 2026  00:41:13 +0000 (0:00:00.166)       0:00:34.806 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'ede169e1-d974-55ed-a938-c303cb89cb5b'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db + wal)] ***********************",
                                "Saturday 21 March 2026  00:41:13 +0000 (0:00:00.163)       0:00:34.970 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'ede169e1-d974-55ed-a938-c303cb89cb5b'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Compile lvm_volumes] *****************************************************",
                                "Saturday 21 March 2026  00:41:13 +0000 (0:00:00.146)       0:00:35.116 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set OSD devices config data] *********************************************",
                                "Saturday 21 March 2026  00:41:13 +0000 (0:00:00.116)       0:00:35.233 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set DB devices config data] **********************************************",
                                "Saturday 21 March 2026  00:41:13 +0000 (0:00:00.115)       0:00:35.348 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set WAL devices config data] *********************************************",
                                "Saturday 21 March 2026  00:41:14 +0000 (0:00:00.103)       0:00:35.452 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set DB+WAL devices config data] ******************************************",
                                "Saturday 21 March 2026  00:41:14 +0000 (0:00:00.119)       0:00:35.571 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print ceph_osd_devices] **************************************************",
                                "Saturday 21 March 2026  00:41:14 +0000 (0:00:00.137)       0:00:35.709 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;32m        \"sdb\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"ede169e1-d974-55ed-a938-c303cb89cb5b\"\u001b[0m",
                                "\u001b[0;32m        },\u001b[0m",
                                "\u001b[0;32m        \"sdc\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"23a81e86-5308-55a4-b4d4-ad2cd0b09f4a\"\u001b[0m",
                                "\u001b[0;32m        }\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print WAL devices] *******************************************************",
                                "Saturday 21 March 2026  00:41:14 +0000 (0:00:00.112)       0:00:35.821 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print DB devices] ********************************************************",
                                "Saturday 21 March 2026  00:41:14 +0000 (0:00:00.095)       0:00:35.917 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print shared DB/WAL devices] *********************************************",
                                "Saturday 21 March 2026  00:41:14 +0000 (0:00:00.266)       0:00:36.183 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print configuration data] ************************************************",
                                "Saturday 21 March 2026  00:41:14 +0000 (0:00:00.116)       0:00:36.300 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m",
                                "\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;33m            \"sdb\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"ede169e1-d974-55ed-a938-c303cb89cb5b\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            \"sdc\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"23a81e86-5308-55a4-b4d4-ad2cd0b09f4a\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        },\u001b[0m",
                                "\u001b[0;33m        \"lvm_volumes\": [\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-ede169e1-d974-55ed-a938-c303cb89cb5b\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        ]\u001b[0m",
                                "\u001b[0;33m    }\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "RUNNING HANDLER [Write configuration file] *************************************",
                                "Saturday 21 March 2026  00:41:15 +0000 (0:00:00.186)       0:00:36.486 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:41:15 +0000 (0:00:00.786)       0:00:37.273 ******** ",
                                "=============================================================================== ",
                                "Write configuration file ------------------------------------------------ 3.93s",
                                "Add known partitions to the list of available block devices ------------- 1.05s",
                                "Add known links to the list of available block devices ------------------ 0.96s",
                                "Get initial list of available block devices ----------------------------- 0.94s",
                                "Get extra vars for Ceph configuration ----------------------------------- 0.83s",
                                "Add known partitions to the list of available block devices ------------- 0.80s",
                                "Add known links to the list of available block devices ------------------ 0.67s",
                                "Add known partitions to the list of available block devices ------------- 0.64s",
                                "Set UUIDs for OSD VGs/LVs ----------------------------------------------- 0.64s",
                                "Generate lvm_volumes structure (block + wal) ---------------------------- 0.61s",
                                "Add known links to the list of available block devices ------------------ 0.58s",
                                "Add known links to the list of available block devices ------------------ 0.57s",
                                "Add known partitions to the list of available block devices ------------- 0.56s",
                                "Print configuration data ------------------------------------------------ 0.56s",
                                "Define lvm_volumes structures ------------------------------------------- 0.54s",
                                "Add known links to the list of available block devices ------------------ 0.53s",
                                "Add known links to the list of available block devices ------------------ 0.53s",
                                "Add known partitions to the list of available block devices ------------- 0.52s",
                                "Add known links to the list of available block devices ------------------ 0.52s",
                                "Add known partitions to the list of available block devices ------------- 0.51s",
                                "\u001b[32m2026-03-21 00:41:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 069d0a36-c9b8-49ce-bd82-37b1b73f47ae (sync inventory) is running in background. Output coming soon.\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting group_vars file reorganization\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMoved 0 file(s) to their respective directories\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGroup_vars file reorganization completed\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting variable preparation from inventory\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-kolla-ceph-rgw-hosts.yml with ceph_rgw_hosts\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-infrastructure-cephclient-mons.yml with cephclient_mons\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-ceph-cluster-fsid.yml with ceph_cluster_fsid\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m3 file(s) written, 6 host(s) processed\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mVariable preparation completed\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting inventory overwrite handling\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mHandling group overwrites in 99-overwrite\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group frr:children from 60-generic\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group netbird:children from 50-infrastructure\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group ceph-mds from 50-ceph\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group ceph-rgw from 50-ceph\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mHandling group overwrites in 20-roles\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group k3s_node from 50-infrastructure\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoved 5 group(s) in total\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mInventory overwrite handling completed\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting merge of inventory files\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mInventory files merged successfully\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGenerating minified hosts file\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSuccessfully wrote minified hosts file to /inventory.merge/hosts-minified.yml\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSuccessfully wrote fast inventory to /inventory.merge/fast/hosts.json\u001b[0m",
                                "\u001b[32m2026-03-21 00:41:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGenerating ClusterShell configuration from Ansible inventory\u001b[0m",
                                "\u001b[32m2026-03-21 00:42:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSuccessfully wrote ClusterShell configuration\u001b[0m",
                                "[master d79765b] 2026-03-21-00-42",
                                " 5 files changed, 75 insertions(+), 10 deletions(-)",
                                " create mode 100644 fast/host_vars/testbed-node-3/ceph-lvm-configuration.yml",
                                " create mode 100644 fast/host_vars/testbed-node-4/ceph-lvm-configuration.yml",
                                " create mode 100644 fast/host_vars/testbed-node-5/ceph-lvm-configuration.yml",
                                "\u001b[32m2026-03-21 00:42:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of ceph-create-lvm-devices.\u001b[0m",
                                "\u001b[32m2026-03-21 00:42:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c172153-90a3-4f8f-a4c7-2223f84e593f (ceph-create-lvm-devices) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 00:42:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 0c172153-90a3-4f8f-a4c7-2223f84e593f (ceph-create-lvm-devices) has been started and output is visible here.\u001b[0m",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Ceph create LVM devices] *************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Saturday 21 March 2026  00:42:14 +0000 (0:00:00.290)       0:00:00.290 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Saturday 21 March 2026  00:42:14 +0000 (0:00:00.236)       0:00:00.527 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:14 +0000 (0:00:00.215)       0:00:00.742 ******** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:15 +0000 (0:00:00.380)       0:00:01.123 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:15 +0000 (0:00:00.458)       0:00:01.582 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:15 +0000 (0:00:00.198)       0:00:01.780 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:16 +0000 (0:00:00.184)       0:00:01.965 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:16 +0000 (0:00:00.190)       0:00:02.155 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:16 +0000 (0:00:00.191)       0:00:02.347 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:16 +0000 (0:00:00.222)       0:00:02.569 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:16 +0000 (0:00:00.204)       0:00:02.773 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:17 +0000 (0:00:00.188)       0:00:02.962 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:17 +0000 (0:00:00.423)       0:00:03.385 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_f08311eb-e2ad-4eb7-8da9-ac6d899a2409)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_f08311eb-e2ad-4eb7-8da9-ac6d899a2409)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:17 +0000 (0:00:00.417)       0:00:03.802 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_e03b171b-b829-41a0-a0b6-cc46d85f013d)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_e03b171b-b829-41a0-a0b6-cc46d85f013d)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:18 +0000 (0:00:00.650)       0:00:04.453 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_9fa3c153-0e2a-45ef-8e48-e0095eb937ad)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_9fa3c153-0e2a-45ef-8e48-e0095eb937ad)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:19 +0000 (0:00:00.625)       0:00:05.079 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:20 +0000 (0:00:00.821)       0:00:05.901 ******** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:20 +0000 (0:00:00.496)       0:00:06.397 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:20 +0000 (0:00:00.218)       0:00:06.616 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:21 +0000 (0:00:00.226)       0:00:06.843 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:21 +0000 (0:00:00.218)       0:00:07.062 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:21 +0000 (0:00:00.219)       0:00:07.281 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:21 +0000 (0:00:00.207)       0:00:07.489 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:21 +0000 (0:00:00.229)       0:00:07.718 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:22 +0000 (0:00:00.228)       0:00:07.947 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:22 +0000 (0:00:00.214)       0:00:08.162 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:23 +0000 (0:00:01.366)       0:00:09.529 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:23 +0000 (0:00:00.212)       0:00:09.741 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:24 +0000 (0:00:00.212)       0:00:09.954 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:24 +0000 (0:00:00.214)       0:00:10.168 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Check whether ceph_db_wal_devices is used exclusively] *******************",
                                "Saturday 21 March 2026  00:42:24 +0000 (0:00:00.232)       0:00:10.401 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************",
                                "Saturday 21 March 2026  00:42:24 +0000 (0:00:00.140)       0:00:10.541 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '64dd100a-6bcc-5763-b557-0732c79bfd55'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '5368e13a-d21f-5912-91bd-da537815d9e8'}})\u001b[0m",
                                "",
                                "TASK [Create block VGs] ********************************************************",
                                "Saturday 21 March 2026  00:42:24 +0000 (0:00:00.224)       0:00:10.765 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block VGs'] ************************************************",
                                "Saturday 21 March 2026  00:42:26 +0000 (0:00:02.026)       0:00:12.792 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create block LVs] ********************************************************",
                                "Saturday 21 March 2026  00:42:27 +0000 (0:00:00.159)       0:00:12.951 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block LVs'] ************************************************",
                                "Saturday 21 March 2026  00:42:28 +0000 (0:00:01.455)       0:00:14.406 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create DB VGs] ***********************************************************",
                                "Saturday 21 March 2026  00:42:28 +0000 (0:00:00.179)       0:00:14.586 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB VGs'] ***************************************************",
                                "Saturday 21 March 2026  00:42:28 +0000 (0:00:00.149)       0:00:14.736 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create WAL VGs] **********************************************************",
                                "Saturday 21 March 2026  00:42:29 +0000 (0:00:00.347)       0:00:15.083 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL VGs'] **************************************************",
                                "Saturday 21 March 2026  00:42:29 +0000 (0:00:00.156)       0:00:15.240 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create DB+WAL VGs] *******************************************************",
                                "Saturday 21 March 2026  00:42:29 +0000 (0:00:00.173)       0:00:15.413 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB+WAL VGs'] ***********************************************",
                                "Saturday 21 March 2026  00:42:29 +0000 (0:00:00.179)       0:00:15.592 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Prepare variables for OSD count check] ***********************************",
                                "Saturday 21 March 2026  00:42:29 +0000 (0:00:00.189)       0:00:15.782 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************",
                                "Saturday 21 March 2026  00:42:30 +0000 (0:00:00.138)       0:00:15.920 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************",
                                "Saturday 21 March 2026  00:42:30 +0000 (0:00:00.210)       0:00:16.131 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************",
                                "Saturday 21 March 2026  00:42:30 +0000 (0:00:00.173)       0:00:16.305 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************",
                                "Saturday 21 March 2026  00:42:30 +0000 (0:00:00.175)       0:00:16.480 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************",
                                "Saturday 21 March 2026  00:42:30 +0000 (0:00:00.146)       0:00:16.626 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************",
                                "Saturday 21 March 2026  00:42:30 +0000 (0:00:00.174)       0:00:16.800 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB VG] ***********************************",
                                "Saturday 21 March 2026  00:42:31 +0000 (0:00:00.153)       0:00:16.954 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per WAL VG] **********************************",
                                "Saturday 21 March 2026  00:42:31 +0000 (0:00:00.510)       0:00:17.464 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB+WAL VG] *******************************",
                                "Saturday 21 March 2026  00:42:31 +0000 (0:00:00.154)       0:00:17.619 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Gather DB VGs with total and available size in bytes] ********************",
                                "Saturday 21 March 2026  00:42:31 +0000 (0:00:00.148)       0:00:17.768 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Gather WAL VGs with total and available size in bytes] *******************",
                                "Saturday 21 March 2026  00:42:32 +0000 (0:00:00.746)       0:00:18.515 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Gather DB+WAL VGs with total and available size in bytes] ****************",
                                "Saturday 21 March 2026  00:42:33 +0000 (0:00:00.490)       0:00:19.005 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************",
                                "Saturday 21 March 2026  00:42:33 +0000 (0:00:00.543)       0:00:19.548 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (without buffer)] *************************************",
                                "Saturday 21 March 2026  00:42:33 +0000 (0:00:00.136)       0:00:19.685 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (with buffer)] ****************************************",
                                "Saturday 21 March 2026  00:42:33 +0000 (0:00:00.120)       0:00:19.805 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print LVM VGs report data] ***********************************************",
                                "Saturday 21 March 2026  00:42:34 +0000 (0:00:00.105)       0:00:19.910 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"vgs_report\": {\u001b[0m",
                                "\u001b[0;32m        \"vg\": []\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print LVM VG sizes] ******************************************************",
                                "Saturday 21 March 2026  00:42:34 +0000 (0:00:00.134)       0:00:20.045 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_db_devices] ************************",
                                "Saturday 21 March 2026  00:42:34 +0000 (0:00:00.149)       0:00:20.194 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_db_devices] ****************************",
                                "Saturday 21 March 2026  00:42:34 +0000 (0:00:00.133)       0:00:20.328 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB LVs on ceph_db_devices > available] *******************",
                                "Saturday 21 March 2026  00:42:34 +0000 (0:00:00.358)       0:00:20.686 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_wal_devices] ***********************",
                                "Saturday 21 March 2026  00:42:35 +0000 (0:00:00.142)       0:00:20.828 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_wal_devices] ***************************",
                                "Saturday 21 March 2026  00:42:35 +0000 (0:00:00.132)       0:00:20.961 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************",
                                "Saturday 21 March 2026  00:42:35 +0000 (0:00:00.132)       0:00:21.094 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************",
                                "Saturday 21 March 2026  00:42:35 +0000 (0:00:00.134)       0:00:21.229 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************",
                                "Saturday 21 March 2026  00:42:35 +0000 (0:00:00.135)       0:00:21.364 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************",
                                "Saturday 21 March 2026  00:42:35 +0000 (0:00:00.141)       0:00:21.506 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************",
                                "Saturday 21 March 2026  00:42:35 +0000 (0:00:00.138)       0:00:21.644 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********",
                                "Saturday 21 March 2026  00:42:35 +0000 (0:00:00.127)       0:00:21.772 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************",
                                "Saturday 21 March 2026  00:42:36 +0000 (0:00:00.131)       0:00:21.904 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************",
                                "Saturday 21 March 2026  00:42:36 +0000 (0:00:00.126)       0:00:22.030 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_devices] ***************************************",
                                "Saturday 21 March 2026  00:42:36 +0000 (0:00:00.139)       0:00:22.169 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_devices'] *******************************",
                                "Saturday 21 March 2026  00:42:36 +0000 (0:00:00.188)       0:00:22.357 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_wal_devices] *************************************",
                                "Saturday 21 March 2026  00:42:36 +0000 (0:00:00.386)       0:00:22.744 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************",
                                "Saturday 21 March 2026  00:42:37 +0000 (0:00:00.157)       0:00:22.902 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_db_wal_devices] **********************************",
                                "Saturday 21 March 2026  00:42:37 +0000 (0:00:00.173)       0:00:23.075 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************",
                                "Saturday 21 March 2026  00:42:37 +0000 (0:00:00.157)       0:00:23.233 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_wal_devices] ***********************************",
                                "Saturday 21 March 2026  00:42:37 +0000 (0:00:00.154)       0:00:23.388 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************",
                                "Saturday 21 March 2026  00:42:37 +0000 (0:00:00.150)       0:00:23.538 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph LVs with associated VGs] ********************************",
                                "Saturday 21 March 2026  00:42:37 +0000 (0:00:00.173)       0:00:23.712 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph PVs with associated VGs] ********************************",
                                "Saturday 21 March 2026  00:42:38 +0000 (0:00:00.500)       0:00:24.213 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************",
                                "Saturday 21 March 2026  00:42:38 +0000 (0:00:00.533)       0:00:24.746 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create list of VG/LV names] **********************************************",
                                "Saturday 21 March 2026  00:42:39 +0000 (0:00:00.151)       0:00:24.898 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'lv_name': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'vg_name': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'lv_name': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'vg_name': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'})\u001b[0m",
                                "",
                                "TASK [Fail if block LV defined in lvm_volumes is missing] **********************",
                                "Saturday 21 March 2026  00:42:39 +0000 (0:00:00.183)       0:00:25.082 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV defined in lvm_volumes is missing] *************************",
                                "Saturday 21 March 2026  00:42:39 +0000 (0:00:00.148)       0:00:25.230 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if WAL LV defined in lvm_volumes is missing] ************************",
                                "Saturday 21 March 2026  00:42:39 +0000 (0:00:00.368)       0:00:25.599 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print LVM report data] ***************************************************",
                                "Saturday 21 March 2026  00:42:39 +0000 (0:00:00.171)       0:00:25.771 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"lvm_report\": {\u001b[0m",
                                "\u001b[0;32m        \"lv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-5368e13a-d21f-5912-91bd-da537815d9e8\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-5368e13a-d21f-5912-91bd-da537815d9e8\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-64dd100a-6bcc-5763-b557-0732c79bfd55\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ],\u001b[0m",
                                "\u001b[0;32m        \"pv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-64dd100a-6bcc-5763-b557-0732c79bfd55\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-5368e13a-d21f-5912-91bd-da537815d9e8\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ]\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY [Ceph create LVM devices] *************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Saturday 21 March 2026  00:42:40 +0000 (0:00:00.304)       0:00:26.076 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Saturday 21 March 2026  00:42:40 +0000 (0:00:00.262)       0:00:26.338 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:40 +0000 (0:00:00.229)       0:00:26.568 ******** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:41 +0000 (0:00:00.387)       0:00:26.956 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:41 +0000 (0:00:00.196)       0:00:27.152 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:41 +0000 (0:00:00.215)       0:00:27.368 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:41 +0000 (0:00:00.226)       0:00:27.595 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:42 +0000 (0:00:00.641)       0:00:28.237 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:42 +0000 (0:00:00.198)       0:00:28.436 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:42 +0000 (0:00:00.222)       0:00:28.658 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:43 +0000 (0:00:00.221)       0:00:28.880 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:43 +0000 (0:00:00.226)       0:00:29.107 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:43 +0000 (0:00:00.467)       0:00:29.574 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_4e74a544-8b3a-4a05-98a1-b7cc13f13c18)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_4e74a544-8b3a-4a05-98a1-b7cc13f13c18)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:44 +0000 (0:00:00.418)       0:00:29.993 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_666e73fe-87f5-4f65-9ff2-95a038c8b18d)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_666e73fe-87f5-4f65-9ff2-95a038c8b18d)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:44 +0000 (0:00:00.384)       0:00:30.378 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_8ae4aabb-0174-4999-a1d0-4bf78cbb2965)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_8ae4aabb-0174-4999-a1d0-4bf78cbb2965)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:42:44 +0000 (0:00:00.390)       0:00:30.769 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:45 +0000 (0:00:00.303)       0:00:31.072 ******** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:45 +0000 (0:00:00.490)       0:00:31.563 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:45 +0000 (0:00:00.167)       0:00:31.730 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:46 +0000 (0:00:00.205)       0:00:31.936 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:46 +0000 (0:00:00.185)       0:00:32.121 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:46 +0000 (0:00:00.187)       0:00:32.309 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:46 +0000 (0:00:00.175)       0:00:32.485 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:46 +0000 (0:00:00.183)       0:00:32.669 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:47 +0000 (0:00:00.175)       0:00:32.844 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:47 +0000 (0:00:00.166)       0:00:33.011 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:47 +0000 (0:00:00.744)       0:00:33.755 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:48 +0000 (0:00:00.179)       0:00:33.935 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:48 +0000 (0:00:00.191)       0:00:34.127 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:42:48 +0000 (0:00:00.470)       0:00:34.597 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Check whether ceph_db_wal_devices is used exclusively] *******************",
                                "Saturday 21 March 2026  00:42:48 +0000 (0:00:00.186)       0:00:34.783 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************",
                                "Saturday 21 March 2026  00:42:49 +0000 (0:00:00.120)       0:00:34.904 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '45b09425-beda-50b4-a83b-1fc50b5adca4'}})\u001b[0m",
                                "",
                                "TASK [Create block VGs] ********************************************************",
                                "Saturday 21 March 2026  00:42:49 +0000 (0:00:00.168)       0:00:35.073 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block VGs'] ************************************************",
                                "Saturday 21 March 2026  00:42:51 +0000 (0:00:01.843)       0:00:36.917 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create block LVs] ********************************************************",
                                "Saturday 21 March 2026  00:42:51 +0000 (0:00:00.130)       0:00:37.047 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block LVs'] ************************************************",
                                "Saturday 21 March 2026  00:42:52 +0000 (0:00:01.193)       0:00:38.240 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create DB VGs] ***********************************************************",
                                "Saturday 21 March 2026  00:42:52 +0000 (0:00:00.124)       0:00:38.365 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB VGs'] ***************************************************",
                                "Saturday 21 March 2026  00:42:52 +0000 (0:00:00.129)       0:00:38.494 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create WAL VGs] **********************************************************",
                                "Saturday 21 March 2026  00:42:52 +0000 (0:00:00.135)       0:00:38.630 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL VGs'] **************************************************",
                                "Saturday 21 March 2026  00:42:52 +0000 (0:00:00.125)       0:00:38.755 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create DB+WAL VGs] *******************************************************",
                                "Saturday 21 March 2026  00:42:53 +0000 (0:00:00.134)       0:00:38.890 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB+WAL VGs'] ***********************************************",
                                "Saturday 21 March 2026  00:42:53 +0000 (0:00:00.258)       0:00:39.148 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Prepare variables for OSD count check] ***********************************",
                                "Saturday 21 March 2026  00:42:53 +0000 (0:00:00.134)       0:00:39.283 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************",
                                "Saturday 21 March 2026  00:42:53 +0000 (0:00:00.125)       0:00:39.408 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************",
                                "Saturday 21 March 2026  00:42:53 +0000 (0:00:00.138)       0:00:39.546 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************",
                                "Saturday 21 March 2026  00:42:53 +0000 (0:00:00.149)       0:00:39.695 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************",
                                "Saturday 21 March 2026  00:42:54 +0000 (0:00:00.140)       0:00:39.836 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************",
                                "Saturday 21 March 2026  00:42:54 +0000 (0:00:00.118)       0:00:39.955 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************",
                                "Saturday 21 March 2026  00:42:54 +0000 (0:00:00.136)       0:00:40.091 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB VG] ***********************************",
                                "Saturday 21 March 2026  00:42:54 +0000 (0:00:00.131)       0:00:40.222 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per WAL VG] **********************************",
                                "Saturday 21 March 2026  00:42:54 +0000 (0:00:00.114)       0:00:40.337 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB+WAL VG] *******************************",
                                "Saturday 21 March 2026  00:42:54 +0000 (0:00:00.129)       0:00:40.467 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Gather DB VGs with total and available size in bytes] ********************",
                                "Saturday 21 March 2026  00:42:54 +0000 (0:00:00.105)       0:00:40.573 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Gather WAL VGs with total and available size in bytes] *******************",
                                "Saturday 21 March 2026  00:42:55 +0000 (0:00:00.692)       0:00:41.266 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Gather DB+WAL VGs with total and available size in bytes] ****************",
                                "Saturday 21 March 2026  00:42:56 +0000 (0:00:00.592)       0:00:41.858 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************",
                                "Saturday 21 March 2026  00:42:56 +0000 (0:00:00.594)       0:00:42.452 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (without buffer)] *************************************",
                                "Saturday 21 March 2026  00:42:56 +0000 (0:00:00.163)       0:00:42.616 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (with buffer)] ****************************************",
                                "Saturday 21 March 2026  00:42:56 +0000 (0:00:00.116)       0:00:42.732 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print LVM VGs report data] ***********************************************",
                                "Saturday 21 March 2026  00:42:57 +0000 (0:00:00.105)       0:00:42.837 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"vgs_report\": {\u001b[0m",
                                "\u001b[0;32m        \"vg\": []\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print LVM VG sizes] ******************************************************",
                                "Saturday 21 March 2026  00:42:57 +0000 (0:00:00.136)       0:00:42.974 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_db_devices] ************************",
                                "Saturday 21 March 2026  00:42:57 +0000 (0:00:00.152)       0:00:43.126 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_db_devices] ****************************",
                                "Saturday 21 March 2026  00:42:57 +0000 (0:00:00.141)       0:00:43.268 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB LVs on ceph_db_devices > available] *******************",
                                "Saturday 21 March 2026  00:42:57 +0000 (0:00:00.133)       0:00:43.402 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_wal_devices] ***********************",
                                "Saturday 21 March 2026  00:42:57 +0000 (0:00:00.134)       0:00:43.537 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_wal_devices] ***************************",
                                "Saturday 21 March 2026  00:42:57 +0000 (0:00:00.137)       0:00:43.674 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************",
                                "Saturday 21 March 2026  00:42:58 +0000 (0:00:00.326)       0:00:44.000 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************",
                                "Saturday 21 March 2026  00:42:58 +0000 (0:00:00.141)       0:00:44.141 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************",
                                "Saturday 21 March 2026  00:42:58 +0000 (0:00:00.134)       0:00:44.276 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************",
                                "Saturday 21 March 2026  00:42:58 +0000 (0:00:00.137)       0:00:44.414 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************",
                                "Saturday 21 March 2026  00:42:58 +0000 (0:00:00.121)       0:00:44.535 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********",
                                "Saturday 21 March 2026  00:42:58 +0000 (0:00:00.148)       0:00:44.683 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************",
                                "Saturday 21 March 2026  00:42:59 +0000 (0:00:00.161)       0:00:44.845 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************",
                                "Saturday 21 March 2026  00:42:59 +0000 (0:00:00.123)       0:00:44.969 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_devices] ***************************************",
                                "Saturday 21 March 2026  00:42:59 +0000 (0:00:00.130)       0:00:45.099 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_devices'] *******************************",
                                "Saturday 21 March 2026  00:42:59 +0000 (0:00:00.165)       0:00:45.264 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_wal_devices] *************************************",
                                "Saturday 21 March 2026  00:42:59 +0000 (0:00:00.158)       0:00:45.422 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************",
                                "Saturday 21 March 2026  00:42:59 +0000 (0:00:00.161)       0:00:45.584 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_db_wal_devices] **********************************",
                                "Saturday 21 March 2026  00:43:00 +0000 (0:00:00.366)       0:00:45.950 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************",
                                "Saturday 21 March 2026  00:43:00 +0000 (0:00:00.172)       0:00:46.123 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_wal_devices] ***********************************",
                                "Saturday 21 March 2026  00:43:00 +0000 (0:00:00.149)       0:00:46.273 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************",
                                "Saturday 21 March 2026  00:43:00 +0000 (0:00:00.134)       0:00:46.407 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph LVs with associated VGs] ********************************",
                                "Saturday 21 March 2026  00:43:00 +0000 (0:00:00.156)       0:00:46.564 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph PVs with associated VGs] ********************************",
                                "Saturday 21 March 2026  00:43:01 +0000 (0:00:00.504)       0:00:47.069 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************",
                                "Saturday 21 March 2026  00:43:01 +0000 (0:00:00.500)       0:00:47.569 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create list of VG/LV names] **********************************************",
                                "Saturday 21 March 2026  00:43:01 +0000 (0:00:00.118)       0:00:47.687 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'lv_name': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'vg_name': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'lv_name': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'vg_name': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'})\u001b[0m",
                                "",
                                "TASK [Fail if block LV defined in lvm_volumes is missing] **********************",
                                "Saturday 21 March 2026  00:43:02 +0000 (0:00:00.151)       0:00:47.839 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV defined in lvm_volumes is missing] *************************",
                                "Saturday 21 March 2026  00:43:02 +0000 (0:00:00.146)       0:00:47.985 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if WAL LV defined in lvm_volumes is missing] ************************",
                                "Saturday 21 March 2026  00:43:02 +0000 (0:00:00.147)       0:00:48.133 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print LVM report data] ***************************************************",
                                "Saturday 21 March 2026  00:43:02 +0000 (0:00:00.158)       0:00:48.291 ******** ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"lvm_report\": {\u001b[0m",
                                "\u001b[0;32m        \"lv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-45b09425-beda-50b4-a83b-1fc50b5adca4\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ],\u001b[0m",
                                "\u001b[0;32m        \"pv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-45b09425-beda-50b4-a83b-1fc50b5adca4\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ]\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY [Ceph create LVM devices] *************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Saturday 21 March 2026  00:43:02 +0000 (0:00:00.401)       0:00:48.692 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Saturday 21 March 2026  00:43:03 +0000 (0:00:00.217)       0:00:48.910 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:43:03 +0000 (0:00:00.226)       0:00:49.137 ******** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:43:03 +0000 (0:00:00.377)       0:00:49.515 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:43:03 +0000 (0:00:00.181)       0:00:49.696 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:43:04 +0000 (0:00:00.182)       0:00:49.878 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:43:04 +0000 (0:00:00.182)       0:00:50.061 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:43:04 +0000 (0:00:00.179)       0:00:50.241 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:43:04 +0000 (0:00:00.183)       0:00:50.424 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:43:05 +0000 (0:00:00.472)       0:00:50.896 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:43:05 +0000 (0:00:00.160)       0:00:51.057 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:43:05 +0000 (0:00:00.165)       0:00:51.223 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:43:05 +0000 (0:00:00.368)       0:00:51.591 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_f76ae8f0-65cc-4a62-b19c-23f8a40fbde6)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_f76ae8f0-65cc-4a62-b19c-23f8a40fbde6)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:43:06 +0000 (0:00:00.372)       0:00:51.964 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_d7bb81ba-6491-4b5a-bb82-f0b49314eac0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_d7bb81ba-6491-4b5a-bb82-f0b49314eac0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:43:06 +0000 (0:00:00.386)       0:00:52.351 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_cdffe5f1-7f0d-460e-a1db-d6846d226144)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_cdffe5f1-7f0d-460e-a1db-d6846d226144)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Saturday 21 March 2026  00:43:06 +0000 (0:00:00.385)       0:00:52.736 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:43:07 +0000 (0:00:00.296)       0:00:53.033 ******** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:43:07 +0000 (0:00:00.396)       0:00:53.429 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:43:07 +0000 (0:00:00.181)       0:00:53.610 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:43:07 +0000 (0:00:00.190)       0:00:53.801 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:43:08 +0000 (0:00:00.565)       0:00:54.366 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:43:08 +0000 (0:00:00.219)       0:00:54.586 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:43:08 +0000 (0:00:00.213)       0:00:54.799 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:43:09 +0000 (0:00:00.211)       0:00:55.011 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:43:09 +0000 (0:00:00.209)       0:00:55.221 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:43:09 +0000 (0:00:00.216)       0:00:55.438 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:43:10 +0000 (0:00:00.641)       0:00:56.080 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:43:10 +0000 (0:00:00.198)       0:00:56.279 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:43:10 +0000 (0:00:00.215)       0:00:56.494 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Saturday 21 March 2026  00:43:10 +0000 (0:00:00.225)       0:00:56.719 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Check whether ceph_db_wal_devices is used exclusively] *******************",
                                "Saturday 21 March 2026  00:43:11 +0000 (0:00:00.220)       0:00:56.940 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************",
                                "Saturday 21 March 2026  00:43:11 +0000 (0:00:00.405)       0:00:57.346 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'ede169e1-d974-55ed-a938-c303cb89cb5b'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}})\u001b[0m",
                                "",
                                "TASK [Create block VGs] ********************************************************",
                                "Saturday 21 March 2026  00:43:11 +0000 (0:00:00.243)       0:00:57.589 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block VGs'] ************************************************",
                                "Saturday 21 March 2026  00:43:13 +0000 (0:00:01.829)       0:00:59.419 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create block LVs] ********************************************************",
                                "Saturday 21 March 2026  00:43:13 +0000 (0:00:00.166)       0:00:59.586 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block LVs'] ************************************************",
                                "Saturday 21 March 2026  00:43:15 +0000 (0:00:01.303)       0:01:00.889 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create DB VGs] ***********************************************************",
                                "Saturday 21 March 2026  00:43:15 +0000 (0:00:00.166)       0:01:01.055 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB VGs'] ***************************************************",
                                "Saturday 21 March 2026  00:43:15 +0000 (0:00:00.129)       0:01:01.185 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create WAL VGs] **********************************************************",
                                "Saturday 21 March 2026  00:43:15 +0000 (0:00:00.152)       0:01:01.338 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL VGs'] **************************************************",
                                "Saturday 21 March 2026  00:43:15 +0000 (0:00:00.135)       0:01:01.473 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create DB+WAL VGs] *******************************************************",
                                "Saturday 21 March 2026  00:43:15 +0000 (0:00:00.141)       0:01:01.614 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB+WAL VGs'] ***********************************************",
                                "Saturday 21 March 2026  00:43:15 +0000 (0:00:00.131)       0:01:01.746 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Prepare variables for OSD count check] ***********************************",
                                "Saturday 21 March 2026  00:43:16 +0000 (0:00:00.157)       0:01:01.904 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************",
                                "Saturday 21 March 2026  00:43:16 +0000 (0:00:00.134)       0:01:02.038 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************",
                                "Saturday 21 March 2026  00:43:16 +0000 (0:00:00.337)       0:01:02.376 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************",
                                "Saturday 21 March 2026  00:43:16 +0000 (0:00:00.161)       0:01:02.537 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************",
                                "Saturday 21 March 2026  00:43:16 +0000 (0:00:00.153)       0:01:02.690 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************",
                                "Saturday 21 March 2026  00:43:17 +0000 (0:00:00.139)       0:01:02.830 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************",
                                "Saturday 21 March 2026  00:43:17 +0000 (0:00:00.138)       0:01:02.968 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB VG] ***********************************",
                                "Saturday 21 March 2026  00:43:17 +0000 (0:00:00.126)       0:01:03.094 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per WAL VG] **********************************",
                                "Saturday 21 March 2026  00:43:17 +0000 (0:00:00.139)       0:01:03.234 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB+WAL VG] *******************************",
                                "Saturday 21 March 2026  00:43:17 +0000 (0:00:00.139)       0:01:03.374 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Gather DB VGs with total and available size in bytes] ********************",
                                "Saturday 21 March 2026  00:43:17 +0000 (0:00:00.142)       0:01:03.516 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Gather WAL VGs with total and available size in bytes] *******************",
                                "Saturday 21 March 2026  00:43:18 +0000 (0:00:00.550)       0:01:04.067 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Gather DB+WAL VGs with total and available size in bytes] ****************",
                                "Saturday 21 March 2026  00:43:18 +0000 (0:00:00.574)       0:01:04.642 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************",
                                "Saturday 21 March 2026  00:43:19 +0000 (0:00:00.509)       0:01:05.151 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (without buffer)] *************************************",
                                "Saturday 21 March 2026  00:43:19 +0000 (0:00:00.337)       0:01:05.489 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (with buffer)] ****************************************",
                                "Saturday 21 March 2026  00:43:19 +0000 (0:00:00.103)       0:01:05.593 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print LVM VGs report data] ***********************************************",
                                "Saturday 21 March 2026  00:43:19 +0000 (0:00:00.106)       0:01:05.699 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"vgs_report\": {\u001b[0m",
                                "\u001b[0;32m        \"vg\": []\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print LVM VG sizes] ******************************************************",
                                "Saturday 21 March 2026  00:43:20 +0000 (0:00:00.139)       0:01:05.838 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_db_devices] ************************",
                                "Saturday 21 March 2026  00:43:20 +0000 (0:00:00.139)       0:01:05.978 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_db_devices] ****************************",
                                "Saturday 21 March 2026  00:43:20 +0000 (0:00:00.133)       0:01:06.112 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB LVs on ceph_db_devices > available] *******************",
                                "Saturday 21 March 2026  00:43:20 +0000 (0:00:00.132)       0:01:06.244 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_wal_devices] ***********************",
                                "Saturday 21 March 2026  00:43:20 +0000 (0:00:00.137)       0:01:06.381 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_wal_devices] ***************************",
                                "Saturday 21 March 2026  00:43:20 +0000 (0:00:00.134)       0:01:06.516 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************",
                                "Saturday 21 March 2026  00:43:20 +0000 (0:00:00.136)       0:01:06.652 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************",
                                "Saturday 21 March 2026  00:43:20 +0000 (0:00:00.110)       0:01:06.762 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************",
                                "Saturday 21 March 2026  00:43:21 +0000 (0:00:00.129)       0:01:06.892 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************",
                                "Saturday 21 March 2026  00:43:21 +0000 (0:00:00.338)       0:01:07.230 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************",
                                "Saturday 21 March 2026  00:43:21 +0000 (0:00:00.139)       0:01:07.370 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********",
                                "Saturday 21 March 2026  00:43:21 +0000 (0:00:00.138)       0:01:07.508 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************",
                                "Saturday 21 March 2026  00:43:21 +0000 (0:00:00.137)       0:01:07.645 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************",
                                "Saturday 21 March 2026  00:43:21 +0000 (0:00:00.145)       0:01:07.791 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_devices] ***************************************",
                                "Saturday 21 March 2026  00:43:22 +0000 (0:00:00.139)       0:01:07.931 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_devices'] *******************************",
                                "Saturday 21 March 2026  00:43:22 +0000 (0:00:00.151)       0:01:08.082 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_wal_devices] *************************************",
                                "Saturday 21 March 2026  00:43:22 +0000 (0:00:00.153)       0:01:08.235 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************",
                                "Saturday 21 March 2026  00:43:22 +0000 (0:00:00.160)       0:01:08.396 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_db_wal_devices] **********************************",
                                "Saturday 21 March 2026  00:43:22 +0000 (0:00:00.146)       0:01:08.543 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************",
                                "Saturday 21 March 2026  00:43:22 +0000 (0:00:00.158)       0:01:08.701 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_wal_devices] ***********************************",
                                "Saturday 21 March 2026  00:43:23 +0000 (0:00:00.153)       0:01:08.854 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************",
                                "Saturday 21 March 2026  00:43:23 +0000 (0:00:00.359)       0:01:09.214 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph LVs with associated VGs] ********************************",
                                "Saturday 21 March 2026  00:43:23 +0000 (0:00:00.160)       0:01:09.374 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph PVs with associated VGs] ********************************",
                                "Saturday 21 March 2026  00:43:24 +0000 (0:00:00.530)       0:01:09.904 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************",
                                "Saturday 21 March 2026  00:43:24 +0000 (0:00:00.569)       0:01:10.473 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create list of VG/LV names] **********************************************",
                                "Saturday 21 March 2026  00:43:24 +0000 (0:00:00.138)       0:01:10.612 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'lv_name': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'vg_name': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'lv_name': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'vg_name': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'})\u001b[0m",
                                "",
                                "TASK [Fail if block LV defined in lvm_volumes is missing] **********************",
                                "Saturday 21 March 2026  00:43:24 +0000 (0:00:00.170)       0:01:10.782 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV defined in lvm_volumes is missing] *************************",
                                "Saturday 21 March 2026  00:43:25 +0000 (0:00:00.155)       0:01:10.937 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if WAL LV defined in lvm_volumes is missing] ************************",
                                "Saturday 21 March 2026  00:43:25 +0000 (0:00:00.157)       0:01:11.095 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print LVM report data] ***************************************************",
                                "Saturday 21 March 2026  00:43:25 +0000 (0:00:00.172)       0:01:11.267 ******** ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"lvm_report\": {\u001b[0m",
                                "\u001b[0;32m        \"lv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-ede169e1-d974-55ed-a938-c303cb89cb5b\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ],\u001b[0m",
                                "\u001b[0;32m        \"pv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-ede169e1-d974-55ed-a938-c303cb89cb5b\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ]\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:43:25 +0000 (0:00:00.145)       0:01:11.413 ******** ",
                                "=============================================================================== ",
                                "Create block VGs -------------------------------------------------------- 5.70s",
                                "Create block LVs -------------------------------------------------------- 3.95s",
                                "Gather DB VGs with total and available size in bytes -------------------- 1.99s",
                                "Gather WAL VGs with total and available size in bytes ------------------- 1.66s",
                                "Gather DB+WAL VGs with total and available size in bytes ---------------- 1.65s",
                                "Get list of Ceph PVs with associated VGs -------------------------------- 1.60s",
                                "Get list of Ceph LVs with associated VGs -------------------------------- 1.54s",
                                "Add known partitions to the list of available block devices ------------- 1.38s",
                                "Add known partitions to the list of available block devices ------------- 1.37s",
                                "Add known links to the list of available block devices ------------------ 1.15s",
                                "Print LVM report data --------------------------------------------------- 0.85s",
                                "Add known links to the list of available block devices ------------------ 0.82s",
                                "Print number of OSDs wanted per DB VG ----------------------------------- 0.76s",
                                "Add known partitions to the list of available block devices ------------- 0.74s",
                                "Get extra vars for Ceph configuration ----------------------------------- 0.72s",
                                "Print 'Create DB LVs for ceph_db_devices' ------------------------------- 0.70s",
                                "Print 'Create WAL LVs for ceph_wal_devices' ----------------------------- 0.69s",
                                "Count OSDs put on ceph_db_devices defined in lvm_volumes ---------------- 0.69s",
                                "Fail if DB LV defined in lvm_volumes is missing ------------------------- 0.67s",
                                "Get initial list of available block devices ----------------------------- 0.67s",
                                "\u001b[32m2026-03-21 00:43:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of facts.\u001b[0m",
                                "\u001b[32m2026-03-21 00:43:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9c028823-76e0-4a79-8ec2-6e01560eaf6e (facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 00:43:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 9c028823-76e0-4a79-8ec2-6e01560eaf6e (facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role facts] ********************************************************",
                                "",
                                "TASK [osism.commons.facts : Create custom facts directory] *********************",
                                "Saturday 21 March 2026  00:43:40 +0000 (0:00:00.328)       0:00:00.328 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Copy fact files] ***********************************",
                                "Saturday 21 March 2026  00:43:42 +0000 (0:00:01.265)       0:00:01.594 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Saturday 21 March 2026  00:43:43 +0000 (0:00:01.063)       0:00:02.657 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts if using --limit] *****************************",
                                "",
                                "TASK [Gather facts for all hosts] **********************************************",
                                "Saturday 21 March 2026  00:43:48 +0000 (0:00:04.876)       0:00:07.534 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:43:48 +0000 (0:00:00.500)       0:00:08.035 ******** ",
                                "=============================================================================== ",
                                "Gathers facts about hosts ----------------------------------------------- 4.88s",
                                "osism.commons.facts : Create custom facts directory --------------------- 1.27s",
                                "osism.commons.facts : Copy fact files ----------------------------------- 1.06s",
                                "Gather facts for all hosts ---------------------------------------------- 0.50s",
                                "\u001b[32m2026-03-21 00:44:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of frr.\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dc4b529e-724d-4a2f-88cc-989a0ee9615c (frr) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task dc4b529e-724d-4a2f-88cc-989a0ee9615c (frr) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role frr] **********************************************************",
                                "",
                                "TASK [osism.services.frr : Include distribution specific install tasks] ********",
                                "Saturday 21 March 2026  00:44:02 +0000 (0:00:00.227)       0:00:00.227 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/frr/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Pin frr package version] ****************************",
                                "Saturday 21 March 2026  00:44:03 +0000 (0:00:00.213)       0:00:00.441 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Install frr package] ********************************",
                                "Saturday 21 March 2026  00:44:04 +0000 (0:00:01.395)       0:00:01.837 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Copy file: /etc/frr/vtysh.conf] *********************",
                                "Saturday 21 March 2026  00:44:13 +0000 (0:00:08.534)       0:00:10.372 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Copy file: /etc/frr/daemons] ************************",
                                "Saturday 21 March 2026  00:44:13 +0000 (0:00:00.944)       0:00:11.316 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Set _frr_uplinks fact] ******************************",
                                "Saturday 21 March 2026  00:44:14 +0000 (0:00:00.829)       0:00:12.146 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Write frr_config_template to temporary file] ********",
                                "Saturday 21 March 2026  00:44:15 +0000 (0:00:01.108)       0:00:13.255 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Render frr.conf from frr_config_template variable] ***",
                                "Saturday 21 March 2026  00:44:16 +0000 (0:00:00.149)       0:00:13.404 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Remove temporary frr_config_template file] **********",
                                "Saturday 21 March 2026  00:44:16 +0000 (0:00:00.255)       0:00:13.659 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Check for frr.conf file in the configuration repository] ***",
                                "Saturday 21 March 2026  00:44:16 +0000 (0:00:00.148)       0:00:13.808 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Copy frr.conf file from the configuration repository] ***",
                                "Saturday 21 March 2026  00:44:16 +0000 (0:00:00.132)       0:00:13.941 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Copy default frr.conf file of type k3s_cilium] ******",
                                "Saturday 21 March 2026  00:44:16 +0000 (0:00:00.143)       0:00:14.084 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Set sysctl parameters] ******************************",
                                "Saturday 21 March 2026  00:44:17 +0000 (0:00:00.893)       0:00:14.978 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.ip_forward', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.send_redirects', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.accept_redirects', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.fib_multipath_hash_policy', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.default.ignore_routes_with_linkdown', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.rp_filter', 'value': 2})\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Manage frr service] *********************************",
                                "Saturday 21 March 2026  00:44:19 +0000 (0:00:02.033)       0:00:17.012 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.frr : Restart frr service] *********************",
                                "Saturday 21 March 2026  00:44:20 +0000 (0:00:01.100)       0:00:18.112 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:44:22 +0000 (0:00:01.315)       0:00:19.428 ******** ",
                                "=============================================================================== ",
                                "osism.services.frr : Install frr package -------------------------------- 8.53s",
                                "osism.services.frr : Set sysctl parameters ------------------------------ 2.03s",
                                "osism.services.frr : Pin frr package version ---------------------------- 1.40s",
                                "osism.services.frr : Restart frr service -------------------------------- 1.32s",
                                "osism.services.frr : Set _frr_uplinks fact ------------------------------ 1.11s",
                                "osism.services.frr : Manage frr service --------------------------------- 1.10s",
                                "osism.services.frr : Copy file: /etc/frr/vtysh.conf --------------------- 0.94s",
                                "osism.services.frr : Copy default frr.conf file of type k3s_cilium ------ 0.89s",
                                "osism.services.frr : Copy file: /etc/frr/daemons ------------------------ 0.83s",
                                "osism.services.frr : Render frr.conf from frr_config_template variable --- 0.26s",
                                "osism.services.frr : Include distribution specific install tasks -------- 0.21s",
                                "osism.services.frr : Write frr_config_template to temporary file -------- 0.15s",
                                "osism.services.frr : Remove temporary frr_config_template file ---------- 0.15s",
                                "osism.services.frr : Copy frr.conf file from the configuration repository --- 0.14s",
                                "osism.services.frr : Check for frr.conf file in the configuration repository --- 0.13s",
                                "",
                                "--> DEPLOY IN A NUTSHELL -- START -- Sat Mar 21 00:44:22 UTC 2026",
                                "",
                                "\u001b[32m2026-03-21 00:44:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCollection nutshell is prepared for execution\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - dotfiles\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - homer\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - netdata\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - openstackclient\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - phpmyadmin\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - common\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- loadbalancer\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- opensearch\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- mariadb-ng\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [3] ---- horizon\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [3] ---- keystone\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- neutron\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ wait-for-nova\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [6] ------- octavia\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- barbican\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- designate\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- ironic\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- placement\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- magnum\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- openvswitch\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- ovn\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- memcached\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- redis\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- rabbitmq-ng\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - kubernetes\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- kubeconfig\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- copy-kubeconfig\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - ceph\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- ceph-pools\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- copy-ceph-keys\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [3] ---- cephclient\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- ceph-bootstrap-dashboard\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- wait-for-keystone\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ kolla-ceph-rgw\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ glance\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ cinder\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ nova\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- prometheus\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ grafana\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAll tasks of the collection nutshell are prepared for execution\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTasks are running in the background\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mNo task IDs specified, wait for all currently running tasks\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask be72019d-4861-4774-8ce4-d34fe16c1c55 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask be72019d-4861-4774-8ce4-d34fe16c1c55 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask be72019d-4861-4774-8ce4-d34fe16c1c55 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask be72019d-4861-4774-8ce4-d34fe16c1c55 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask be72019d-4861-4774-8ce4-d34fe16c1c55 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask be72019d-4861-4774-8ce4-d34fe16c1c55 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask be72019d-4861-4774-8ce4-d34fe16c1c55 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "PLAY [Apply role geerlingguy.dotfiles] *****************************************",
                                "",
                                "TASK [geerlingguy.dotfiles : Ensure dotfiles repository is cloned locally.] ****",
                                "Saturday 21 March 2026  00:44:44 +0000 (0:00:00.673)       0:00:00.673 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [geerlingguy.dotfiles : Ensure all configured dotfiles are links.] ********",
                                "Saturday 21 March 2026  00:44:49 +0000 (0:00:05.308)       0:00:05.982 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=.tmux.conf)\u001b[0m",
                                "",
                                "TASK [geerlingguy.dotfiles : Remove existing dotfiles file if a replacement is being linked.] ***",
                                "Saturday 21 March 2026  00:44:52 +0000 (0:00:02.156)       0:00:08.138 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-21 00:44:50.481228', 'end': '2026-03-21 00:44:51.488814', 'delta': '0:00:01.007586', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-21 00:44:50.685102', 'end': '2026-03-21 00:44:50.691209', 'delta': '0:00:00.006107', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-21 00:44:50.266670', 'end': '2026-03-21 00:44:50.270757', 'delta': '0:00:00.004087', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-21 00:44:50.972699', 'end': '2026-03-21 00:44:50.979364', 'delta': '0:00:00.006665', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-21 00:44:51.476165', 'end': '2026-03-21 00:44:51.485035', 'delta': '0:00:00.008870', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-21 00:44:51.684500', 'end': '2026-03-21 00:44:51.690397', 'delta': '0:00:00.005897', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-21 00:44:50.964795', 'end': '2026-03-21 00:44:50.970733', 'delta': '0:00:00.005938', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "",
                                "TASK [geerlingguy.dotfiles : Ensure parent folders of link dotfiles exist.] ****",
                                "Saturday 21 March 2026  00:44:53 +0000 (0:00:01.376)       0:00:09.515 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=.tmux.conf)\u001b[0m",
                                "",
                                "TASK [geerlingguy.dotfiles : Link dotfiles into home folder.] ******************",
                                "Saturday 21 March 2026  00:44:54 +0000 (0:00:01.332)       0:00:10.847 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=.tmux.conf)\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:44:57 +0000 (0:00:02.576)       0:00:13.424 ******** ",
                                "=============================================================================== ",
                                "geerlingguy.dotfiles : Ensure dotfiles repository is cloned locally. ---- 5.31s",
                                "geerlingguy.dotfiles : Link dotfiles into home folder. ------------------ 2.58s",
                                "geerlingguy.dotfiles : Ensure all configured dotfiles are links. -------- 2.16s",
                                "geerlingguy.dotfiles : Remove existing dotfiles file if a replacement is being linked. --- 1.38s",
                                "geerlingguy.dotfiles : Ensure parent folders of link dotfiles exist. ---- 1.33s",
                                "\u001b[32m2026-03-21 00:44:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask be72019d-4861-4774-8ce4-d34fe16c1c55 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:44:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8f5910e-22c4-4469-ac07-3f511ea31a23 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3f2dc51c-d510-477b-b94c-50f74b45b300 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:45:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87a25819-e174-4970-9b7a-4e543b07766e is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Apply role homer] ********************************************************",
                                "",
                                "TASK [osism.services.homer : Inform about new parameter homer_url_opensearch_dashboards] ***",
                                "Saturday 21 March 2026  00:44:43 +0000 (0:00:00.884)       0:00:00.884 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"The support for the homer_url_kibana has been removed. Please use the homer_url_opensearch_dashboards parameter.\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [osism.services.homer : Create traefik external network] ******************",
                                "Saturday 21 March 2026  00:44:44 +0000 (0:00:00.369)       0:00:01.253 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.homer : Create required directories] **********************",
                                "Saturday 21 March 2026  00:44:47 +0000 (0:00:03.616)       0:00:04.870 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/homer/configuration)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/homer)\u001b[0m",
                                "",
                                "TASK [osism.services.homer : Copy config.yml configuration file] ***************",
                                "Saturday 21 March 2026  00:44:49 +0000 (0:00:01.718)       0:00:06.588 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.homer : Copy docker-compose.yml file] *********************",
                                "Saturday 21 March 2026  00:44:51 +0000 (0:00:02.319)       0:00:08.907 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.homer : Manage homer service] *****************************",
                                "Saturday 21 March 2026  00:44:53 +0000 (0:00:01.883)       0:00:10.792 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage homer service (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.homer : Restart homer service] *****************",
                                "Saturday 21 March 2026  00:45:21 +0000 (0:00:28.166)       0:00:38.958 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:45:27 +0000 (0:00:05.262)       0:00:44.221 ******** ",
                                "=============================================================================== ",
                                "osism.services.homer : Manage homer service ---------------------------- 28.17s",
                                "osism.services.homer : Restart homer service ---------------------------- 5.26s",
                                "osism.services.homer : Create traefik external network ------------------ 3.62s",
                                "osism.services.homer : Copy config.yml configuration file --------------- 2.32s",
                                "osism.services.homer : Copy docker-compose.yml file --------------------- 1.88s",
                                "osism.services.homer : Create required directories ---------------------- 1.72s",
                                "osism.services.homer : Inform about new parameter homer_url_opensearch_dashboards --- 0.37s",
                                "",
                                "",
                                "PLAY [Apply role openstackclient] **********************************************",
                                "",
                                "TASK [osism.services.openstackclient : Include tasks] **************************",
                                "Saturday 21 March 2026  00:44:44 +0000 (0:00:01.129)       0:00:01.129 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/openstackclient/tasks/container-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Create required directories] ************",
                                "Saturday 21 March 2026  00:44:45 +0000 (0:00:00.769)       0:00:01.899 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/configuration/environments/openstack)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/openstackclient/data)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/openstackclient)\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Copy docker-compose.yml file] ***********",
                                "Saturday 21 March 2026  00:44:48 +0000 (0:00:02.972)       0:00:04.872 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Manage openstackclient service] *********",
                                "Saturday 21 March 2026  00:44:50 +0000 (0:00:02.103)       0:00:06.976 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage openstackclient service (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Copy openstack wrapper script] **********",
                                "Saturday 21 March 2026  00:45:25 +0000 (0:00:35.346)       0:00:42.322 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Remove ospurge wrapper script] **********",
                                "Saturday 21 March 2026  00:45:26 +0000 (0:00:00.995)       0:00:43.318 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.openstackclient : Restart openstackclient service] ***",
                                "Saturday 21 March 2026  00:45:27 +0000 (0:00:01.216)       0:00:44.535 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.openstackclient : Ensure that all containers are up] ***",
                                "Saturday 21 March 2026  00:45:29 +0000 (0:00:01.654)       0:00:46.189 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.openstackclient : Wait for an healthy service] ***",
                                "Saturday 21 March 2026  00:45:30 +0000 (0:00:01.476)       0:00:47.665 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.openstackclient : Copy bash completion script] ***",
                                "Saturday 21 March 2026  00:45:31 +0000 (0:00:00.887)       0:00:48.553 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:45:32 +0000 (0:00:00.512)       0:00:49.066 ******** ",
                                "=============================================================================== ",
                                "osism.services.openstackclient : Manage openstackclient service -------- 35.35s",
                                "osism.services.openstackclient : Create required directories ------------ 2.97s",
                                "osism.services.openstackclient : Copy docker-compose.yml file ----------- 2.10s",
                                "osism.services.openstackclient : Restart openstackclient service -------- 1.65s",
                                "osism.services.openstackclient : Ensure that all containers are up ------ 1.48s",
                                "osism.services.openstackclient : Remove ospurge wrapper script ---------- 1.22s",
                                "osism.services.openstackclient : Copy openstack wrapper script ---------- 1.00s",
                                "osism.services.openstackclient : Wait for an healthy service ------------ 0.89s",
                                "osism.services.openstackclient : Include tasks -------------------------- 0.77s",
                                "osism.services.openstackclient : Copy bash completion script ------------ 0.51s",
                                "",
                                "",
                                "PLAY [Apply role phpmyadmin] ***************************************************",
                                "",
                                "TASK [osism.services.phpmyadmin : Create traefik external network] *************",
                                "Saturday 21 March 2026  00:45:05 +0000 (0:00:00.405)       0:00:00.405 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.phpmyadmin : Create required directories] *****************",
                                "Saturday 21 March 2026  00:45:07 +0000 (0:00:01.492)       0:00:01.897 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/phpmyadmin)\u001b[0m",
                                "",
                                "TASK [osism.services.phpmyadmin : Copy docker-compose.yml file] ****************",
                                "Saturday 21 March 2026  00:45:08 +0000 (0:00:00.765)       0:00:02.663 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.phpmyadmin : Manage phpmyadmin service] *******************",
                                "Saturday 21 March 2026  00:45:09 +0000 (0:00:01.645)       0:00:04.309 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage phpmyadmin service (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.phpmyadmin : Restart phpmyadmin service] *******",
                                "Saturday 21 March 2026  00:46:09 +0000 (0:00:59.909)       0:01:04.218 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:46:12 +0000 (0:00:02.982)       0:01:07.201 ******** ",
                                "=============================================================================== ",
                                "osism.services.phpmyadmin : Manage phpmyadmin service ------------------ 59.91s",
                                "osism.services.phpmyadmin : Restart phpmyadmin service ------------------ 2.98s",
                                "osism.services.phpmyadmin : Copy docker-compose.yml file ---------------- 1.65s",
                                "osism.services.phpmyadmin : Create traefik external network ------------- 1.49s",
                                "osism.services.phpmyadmin : Create required directories ----------------- 0.77s",
                                "\u001b[32m2026-03-21 00:46:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  00:44:44 +0000 (0:00:00.844)       0:00:00.844 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=enable_netdata_True)\u001b[0m",
                                "",
                                "PLAY [Apply role netdata] ******************************************************",
                                "",
                                "TASK [osism.services.netdata : Include distribution specific install tasks] ****",
                                "Saturday 21 March 2026  00:44:46 +0000 (0:00:01.737)       0:00:02.581 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Remove old architecture-dependent repository] ***",
                                "Saturday 21 March 2026  00:44:48 +0000 (0:00:02.214)       0:00:04.796 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Install apt-transport-https package] ************",
                                "Saturday 21 March 2026  00:44:50 +0000 (0:00:02.606)       0:00:07.402 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Add repository gpg key] *************************",
                                "Saturday 21 March 2026  00:44:55 +0000 (0:00:04.300)       0:00:11.702 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Add repository] *********************************",
                                "Saturday 21 March 2026  00:44:57 +0000 (0:00:02.212)       0:00:13.915 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Install package netdata] ************************",
                                "Saturday 21 March 2026  00:45:08 +0000 (0:00:10.636)       0:00:24.552 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Include config tasks] ***************************",
                                "Saturday 21 March 2026  00:45:44 +0000 (0:00:36.554)       0:01:01.107 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/config.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Copy configuration files] ***********************",
                                "Saturday 21 March 2026  00:45:46 +0000 (0:00:02.024)       0:01:03.131 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=stream.conf)\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Retrieve /etc/netdata/.opt-out-from-anonymous-statistics status] ***",
                                "Saturday 21 March 2026  00:45:52 +0000 (0:00:06.145)       0:01:09.277 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Opt out from anonymous statistics] **************",
                                "Saturday 21 March 2026  00:45:54 +0000 (0:00:01.470)       0:01:10.747 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Add netdata user to docker group] ***************",
                                "Saturday 21 March 2026  00:45:56 +0000 (0:00:01.899)       0:01:12.647 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Manage service netdata] *************************",
                                "Saturday 21 March 2026  00:45:57 +0000 (0:00:01.641)       0:01:14.289 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Include host type specific tasks] ***************",
                                "Saturday 21 March 2026  00:45:59 +0000 (0:00:01.904)       0:01:16.193 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/server.yml for testbed-manager\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/client.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Set sysctl vm.max_map_count parameter] **********",
                                "Saturday 21 March 2026  00:46:01 +0000 (0:00:01.416)       0:01:17.610 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.netdata : Restart service netdata] *************",
                                "Saturday 21 March 2026  00:46:03 +0000 (0:00:02.636)       0:01:20.246 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:46:15 +0000 (0:00:11.497)       0:01:31.743 ******** ",
                                "=============================================================================== ",
                                "osism.services.netdata : Install package netdata ----------------------- 36.55s",
                                "osism.services.netdata : Restart service netdata ----------------------- 11.50s",
                                "osism.services.netdata : Add repository -------------------------------- 10.64s",
                                "osism.services.netdata : Copy configuration files ----------------------- 6.15s",
                                "osism.services.netdata : Install apt-transport-https package ------------ 4.30s",
                                "osism.services.netdata : Set sysctl vm.max_map_count parameter ---------- 2.64s",
                                "osism.services.netdata : Remove old architecture-dependent repository --- 2.61s",
                                "osism.services.netdata : Include distribution specific install tasks ---- 2.21s",
                                "osism.services.netdata : Add repository gpg key ------------------------- 2.21s",
                                "osism.services.netdata : Include config tasks --------------------------- 2.02s",
                                "osism.services.netdata : Manage service netdata ------------------------- 1.90s",
                                "osism.services.netdata : Opt out from anonymous statistics -------------- 1.90s",
                                "Group hosts based on enabled services ----------------------------------- 1.74s",
                                "osism.services.netdata : Add netdata user to docker group --------------- 1.64s",
                                "osism.services.netdata : Retrieve /etc/netdata/.opt-out-from-anonymous-statistics status --- 1.47s",
                                "osism.services.netdata : Include host type specific tasks --------------- 1.42s",
                                "\u001b[32m2026-03-21 00:46:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1f1d05f4-7586-472f-baec-72289ef2a08d is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:46:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c076440c-db82-4d81-8e70-68ed60306a74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8fd80eee-9c08-4580-b485-c88b841b0e3e is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Apply role common] *******************************************************",
                                "",
                                "TASK [common : include_tasks] **************************************************",
                                "Saturday 21 March 2026  00:44:37 +0000 (0:00:00.282)       0:00:00.282 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/common/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [common : Ensuring config directories exist] ******************************",
                                "Saturday 21 March 2026  00:44:38 +0000 (0:00:01.310)       0:00:01.593 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "",
                                "TASK [common : include_tasks] **************************************************",
                                "Saturday 21 March 2026  00:44:42 +0000 (0:00:03.850)       0:00:05.443 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/common/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [service-cert-copy : common | Copying over extra CA certificates] *********",
                                "Saturday 21 March 2026  00:44:44 +0000 (0:00:01.416)       0:00:06.860 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : common | Copying over backend internal TLS certificate] ***",
                                "Saturday 21 March 2026  00:44:48 +0000 (0:00:04.699)       0:00:11.559 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : common | Copying over backend internal TLS key] ******",
                                "Saturday 21 March 2026  00:44:50 +0000 (0:00:02.052)       0:00:13.612 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Copying over /run subdirectories conf] **************************",
                                "Saturday 21 March 2026  00:44:53 +0000 (0:00:02.440)       0:00:16.053 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Restart systemd-tmpfiles] ***************************************",
                                "Saturday 21 March 2026  00:44:54 +0000 (0:00:01.227)       0:00:17.280 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Copying over config.json files for services] ********************",
                                "Saturday 21 March 2026  00:44:55 +0000 (0:00:01.368)       0:00:18.648 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd input config files] *************************",
                                "Saturday 21 March 2026  00:45:04 +0000 (0:00:08.764)       0:00:27.413 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd filter config files] ************************",
                                "Saturday 21 March 2026  00:45:06 +0000 (0:00:01.451)       0:00:28.864 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd format config files] ************************",
                                "Saturday 21 March 2026  00:45:07 +0000 (0:00:00.904)       0:00:29.768 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd output config files] ************************",
                                "Saturday 21 March 2026  00:45:07 +0000 (0:00:00.869)       0:00:30.637 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Copying over fluentd.conf] **************************************",
                                "Saturday 21 March 2026  00:45:08 +0000 (0:00:00.942)       0:00:31.580 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [common : Copying over cron logrotate config file] ************************",
                                "Saturday 21 March 2026  00:45:13 +0000 (0:00:04.842)       0:00:36.422 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "",
                                "TASK [common : Ensure RabbitMQ Erlang cookie exists] ***************************",
                                "Saturday 21 March 2026  00:45:17 +0000 (0:00:03.422)       0:00:39.845 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Ensuring config directories have correct owner and permission] ***",
                                "Saturday 21 March 2026  00:45:19 +0000 (0:00:02.339)       0:00:42.184 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [common : Copy rabbitmq-env.conf to kolla toolbox] ************************",
                                "Saturday 21 March 2026  00:45:21 +0000 (0:00:02.295)       0:00:44.480 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "",
                                "TASK [common : Copy rabbitmq erl_inetrc to kolla toolbox] **********************",
                                "Saturday 21 March 2026  00:45:26 +0000 (0:00:04.327)       0:00:48.808 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "",
                                "TASK [common : Check common containers] ****************************************",
                                "Saturday 21 March 2026  00:45:29 +0000 (0:00:03.615)       0:00:52.424 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [common : Creating log volume] ********************************************",
                                "Saturday 21 March 2026  00:45:33 +0000 (0:00:03.671)       0:00:56.096 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Link kolla_logs volume to /var/log/kolla] ***********************",
                                "Saturday 21 March 2026  00:45:35 +0000 (0:00:01.766)       0:00:57.862 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Saturday 21 March 2026  00:45:36 +0000 (0:00:01.511)       0:00:59.374 ******** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Saturday 21 March 2026  00:45:36 +0000 (0:00:00.073)       0:00:59.447 ******** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Saturday 21 March 2026  00:45:36 +0000 (0:00:00.067)       0:00:59.514 ******** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Saturday 21 March 2026  00:45:36 +0000 (0:00:00.075)       0:00:59.590 ******** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Saturday 21 March 2026  00:45:36 +0000 (0:00:00.083)       0:00:59.673 ******** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Saturday 21 March 2026  00:45:37 +0000 (0:00:00.068)       0:00:59.742 ******** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Saturday 21 March 2026  00:45:37 +0000 (0:00:00.076)       0:00:59.818 ******** ",
                                "",
                                "RUNNING HANDLER [common : Restart fluentd container] ***************************",
                                "Saturday 21 March 2026  00:45:37 +0000 (0:00:00.092)       0:00:59.910 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [common : Restart kolla-toolbox container] *********************",
                                "Saturday 21 March 2026  00:46:12 +0000 (0:00:34.955)       0:01:34.866 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [common : Initializing toolbox container using normal user] ****",
                                "Saturday 21 March 2026  00:47:03 +0000 (0:00:51.027)       0:02:25.894 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [common : Restart cron container] ******************************",
                                "Saturday 21 March 2026  00:47:04 +0000 (0:00:01.779)       0:02:27.673 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:47:14 +0000 (0:00:09.344)       0:02:37.018 ******** ",
                                "=============================================================================== ",
                                "common : Restart kolla-toolbox container ------------------------------- 51.03s",
                                "common : Restart fluentd container ------------------------------------- 34.96s",
                                "common : Restart cron container ----------------------------------------- 9.34s",
                                "common : Copying over config.json files for services -------------------- 8.76s",
                                "common : Copying over fluentd.conf -------------------------------------- 4.84s",
                                "service-cert-copy : common | Copying over extra CA certificates --------- 4.70s",
                                "common : Copy rabbitmq-env.conf to kolla toolbox ------------------------ 4.33s",
                                "common : Ensuring config directories exist ------------------------------ 3.85s",
                                "common : Check common containers ---------------------------------------- 3.67s",
                                "common : Copy rabbitmq erl_inetrc to kolla toolbox ---------------------- 3.62s",
                                "common : Copying over cron logrotate config file ------------------------ 3.42s",
                                "service-cert-copy : common | Copying over backend internal TLS key ------ 2.44s",
                                "common : Ensure RabbitMQ Erlang cookie exists --------------------------- 2.34s",
                                "common : Ensuring config directories have correct owner and permission --- 2.30s",
                                "service-cert-copy : common | Copying over backend internal TLS certificate --- 2.05s",
                                "common : Initializing toolbox container using normal user --------------- 1.78s",
                                "common : Creating log volume -------------------------------------------- 1.77s",
                                "common : Link kolla_logs volume to /var/log/kolla ----------------------- 1.51s",
                                "common : Find custom fluentd input config files ------------------------- 1.45s",
                                "common : include_tasks -------------------------------------------------- 1.42s",
                                "\u001b[32m2026-03-21 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0d652c86-9d68-4200-b5b9-a2d094b18b9c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c076440c-db82-4d81-8e70-68ed60306a74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0d652c86-9d68-4200-b5b9-a2d094b18b9c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c076440c-db82-4d81-8e70-68ed60306a74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0d652c86-9d68-4200-b5b9-a2d094b18b9c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c076440c-db82-4d81-8e70-68ed60306a74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0d652c86-9d68-4200-b5b9-a2d094b18b9c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c076440c-db82-4d81-8e70-68ed60306a74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0d652c86-9d68-4200-b5b9-a2d094b18b9c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c076440c-db82-4d81-8e70-68ed60306a74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0d652c86-9d68-4200-b5b9-a2d094b18b9c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c076440c-db82-4d81-8e70-68ed60306a74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0d652c86-9d68-4200-b5b9-a2d094b18b9c is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c076440c-db82-4d81-8e70-68ed60306a74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c076440c-db82-4d81-8e70-68ed60306a74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c076440c-db82-4d81-8e70-68ed60306a74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c076440c-db82-4d81-8e70-68ed60306a74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  00:47:19 +0000 (0:00:00.552)       0:00:00.552 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  00:47:20 +0000 (0:00:00.643)       0:00:01.196 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_memcached_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_memcached_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_memcached_True)\u001b[0m",
                                "",
                                "PLAY [Apply role memcached] ****************************************************",
                                "",
                                "TASK [memcached : include_tasks] ***********************************************",
                                "Saturday 21 March 2026  00:47:20 +0000 (0:00:00.477)       0:00:01.673 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/memcached/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [memcached : Ensuring config directories exist] ***************************",
                                "Saturday 21 March 2026  00:47:21 +0000 (0:00:00.531)       0:00:02.205 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=memcached)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=memcached)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=memcached)\u001b[0m",
                                "",
                                "TASK [memcached : Copying over config.json files for services] *****************",
                                "Saturday 21 March 2026  00:47:22 +0000 (0:00:01.730)       0:00:03.935 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=memcached)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=memcached)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=memcached)\u001b[0m",
                                "",
                                "TASK [memcached : Check memcached container] ***********************************",
                                "Saturday 21 March 2026  00:47:25 +0000 (0:00:02.216)       0:00:06.151 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [memcached : Restart memcached container] **********************",
                                "Saturday 21 March 2026  00:47:27 +0000 (0:00:02.571)       0:00:08.722 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:47:34 +0000 (0:00:06.581)       0:00:15.304 ******** ",
                                "=============================================================================== ",
                                "memcached : Restart memcached container --------------------------------- 6.58s",
                                "memcached : Check memcached container ----------------------------------- 2.57s",
                                "memcached : Copying over config.json files for services ----------------- 2.22s",
                                "memcached : Ensuring config directories exist --------------------------- 1.73s",
                                "Group hosts based on Kolla action --------------------------------------- 0.64s",
                                "memcached : include_tasks ----------------------------------------------- 0.53s",
                                "Group hosts based on enabled services ----------------------------------- 0.48s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  00:47:19 +0000 (0:00:00.473)       0:00:00.473 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  00:47:20 +0000 (0:00:00.550)       0:00:01.023 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_redis_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_redis_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_redis_True)\u001b[0m",
                                "",
                                "PLAY [Apply role redis] ********************************************************",
                                "",
                                "TASK [redis : include_tasks] ***************************************************",
                                "Saturday 21 March 2026  00:47:20 +0000 (0:00:00.827)       0:00:01.850 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/redis/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [redis : Ensuring config directories exist] *******************************",
                                "Saturday 21 March 2026  00:47:22 +0000 (0:00:01.288)       0:00:03.139 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [redis : Copying over default config.json files] **************************",
                                "Saturday 21 March 2026  00:47:25 +0000 (0:00:02.777)       0:00:05.916 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [redis : Copying over redis config files] *********************************",
                                "Saturday 21 March 2026  00:47:28 +0000 (0:00:03.126)       0:00:09.043 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [redis : Check redis containers] ******************************************",
                                "Saturday 21 March 2026  00:47:30 +0000 (0:00:02.760)       0:00:11.804 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [redis : Flush handlers] **************************************************",
                                "Saturday 21 March 2026  00:47:32 +0000 (0:00:01.625)       0:00:13.429 ******** ",
                                "",
                                "TASK [redis : Flush handlers] **************************************************",
                                "Saturday 21 March 2026  00:47:32 +0000 (0:00:00.222)       0:00:13.651 ******** ",
                                "",
                                "TASK [redis : Flush handlers] **************************************************",
                                "Saturday 21 March 2026  00:47:32 +0000 (0:00:00.055)       0:00:13.707 ******** ",
                                "",
                                "RUNNING HANDLER [redis : Restart redis container] ******************************",
                                "Saturday 21 March 2026  00:47:32 +0000 (0:00:00.117)       0:00:13.825 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [redis : Restart redis-sentinel container] *********************",
                                "Saturday 21 March 2026  00:47:40 +0000 (0:00:07.616)       0:00:21.441 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:47:50 +0000 (0:00:09.991)       0:00:31.433 ******** ",
                                "=============================================================================== ",
                                "redis : Restart redis-sentinel container -------------------------------- 9.99s",
                                "redis : Restart redis container ----------------------------------------- 7.62s",
                                "redis : Copying over default config.json files -------------------------- 3.13s",
                                "redis : Ensuring config directories exist ------------------------------- 2.78s",
                                "redis : Copying over redis config files --------------------------------- 2.76s",
                                "redis : Check redis containers ------------------------------------------ 1.63s",
                                "redis : include_tasks --------------------------------------------------- 1.29s",
                                "Group hosts based on enabled services ----------------------------------- 0.83s",
                                "Group hosts based on Kolla action --------------------------------------- 0.55s",
                                "redis : Flush handlers -------------------------------------------------- 0.40s",
                                "\u001b[32m2026-03-21 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c076440c-db82-4d81-8e70-68ed60306a74 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:47:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  00:47:19 +0000 (0:00:00.380)       0:00:00.380 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  00:47:20 +0000 (0:00:01.067)       0:00:01.448 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "",
                                "PLAY [Apply role openvswitch] **************************************************",
                                "",
                                "TASK [openvswitch : include_tasks] *********************************************",
                                "Saturday 21 March 2026  00:47:21 +0000 (0:00:01.502)       0:00:02.951 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/openvswitch/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [module-load : Load modules] **********************************************",
                                "Saturday 21 March 2026  00:47:24 +0000 (0:00:02.286)       0:00:05.237 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=openvswitch)\u001b[0m",
                                "",
                                "TASK [module-load : Persist modules via modules-load.d] ************************",
                                "Saturday 21 March 2026  00:47:26 +0000 (0:00:02.369)       0:00:07.607 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=openvswitch)\u001b[0m",
                                "",
                                "TASK [module-load : Drop module persistence] ***********************************",
                                "Saturday 21 March 2026  00:47:28 +0000 (0:00:01.681)       0:00:09.288 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [openvswitch : Create /run/openvswitch directory on host] *****************",
                                "Saturday 21 March 2026  00:47:29 +0000 (0:00:01.706)       0:00:10.995 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [openvswitch : Ensuring config directories exist] *************************",
                                "Saturday 21 March 2026  00:47:30 +0000 (0:00:00.733)       0:00:11.728 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [openvswitch : Copying over config.json files for services] ***************",
                                "Saturday 21 March 2026  00:47:32 +0000 (0:00:01.552)       0:00:13.281 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [openvswitch : Copying over ovs-vsctl wrapper] ****************************",
                                "Saturday 21 March 2026  00:47:34 +0000 (0:00:02.836)       0:00:16.118 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [openvswitch : Check openvswitch containers] ******************************",
                                "Saturday 21 March 2026  00:47:36 +0000 (0:00:01.424)       0:00:17.543 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Saturday 21 March 2026  00:47:39 +0000 (0:00:03.181)       0:00:20.724 ******** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Saturday 21 March 2026  00:47:39 +0000 (0:00:00.193)       0:00:20.918 ******** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Saturday 21 March 2026  00:47:39 +0000 (0:00:00.222)       0:00:21.141 ******** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Saturday 21 March 2026  00:47:40 +0000 (0:00:00.170)       0:00:21.312 ******** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Saturday 21 March 2026  00:47:40 +0000 (0:00:00.324)       0:00:21.636 ******** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Saturday 21 March 2026  00:47:40 +0000 (0:00:00.275)       0:00:21.911 ******** ",
                                "",
                                "RUNNING HANDLER [openvswitch : Restart openvswitch-db-server container] ********",
                                "Saturday 21 March 2026  00:47:41 +0000 (0:00:00.494)       0:00:22.406 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [openvswitch : Waiting for openvswitch_db service to be ready] ***",
                                "Saturday 21 March 2026  00:47:51 +0000 (0:00:10.009)       0:00:32.416 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [openvswitch : Restart openvswitch-vswitchd container] *********",
                                "Saturday 21 March 2026  00:47:52 +0000 (0:00:01.554)       0:00:33.971 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [openvswitch : Set system-id, hostname and hw-offload] ********************",
                                "Saturday 21 March 2026  00:48:03 +0000 (0:00:10.281)       0:00:44.252 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-3'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-4'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-1'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-5'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-0'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-2'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-3'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-1'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-4'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-0'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-5'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-2'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "",
                                "TASK [openvswitch : Ensuring OVS bridge is properly setup] *********************",
                                "Saturday 21 March 2026  00:48:10 +0000 (0:00:07.157)       0:00:51.409 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=br-ex) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=br-ex) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=br-ex) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=br-ex)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=br-ex)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=br-ex)\u001b[0m",
                                "",
                                "TASK [openvswitch : Ensuring OVS ports are properly setup] *********************",
                                "Saturday 21 March 2026  00:48:13 +0000 (0:00:02.927)       0:00:54.336 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=['br-ex', 'vxlan0']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=['br-ex', 'vxlan0']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=['br-ex', 'vxlan0']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=['br-ex', 'vxlan0'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=['br-ex', 'vxlan0'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=['br-ex', 'vxlan0'])\u001b[0m",
                                "",
                                "RUNNING HANDLER [openvswitch : Restart openvswitch-vswitchd container] *********",
                                "Saturday 21 March 2026  00:48:16 +0000 (0:00:03.512)       0:00:57.849 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:48:24 +0000 (0:00:08.015)       0:01:05.865 ******** ",
                                "=============================================================================== ",
                                "openvswitch : Restart openvswitch-vswitchd container ------------------- 18.30s",
                                "openvswitch : Restart openvswitch-db-server container ------------------ 10.01s",
                                "openvswitch : Set system-id, hostname and hw-offload -------------------- 7.16s",
                                "openvswitch : Ensuring OVS ports are properly setup --------------------- 3.51s",
                                "openvswitch : Check openvswitch containers ------------------------------ 3.18s",
                                "openvswitch : Ensuring OVS bridge is properly setup --------------------- 2.93s",
                                "openvswitch : Copying over config.json files for services --------------- 2.84s",
                                "module-load : Load modules ---------------------------------------------- 2.37s",
                                "openvswitch : include_tasks --------------------------------------------- 2.29s",
                                "module-load : Drop module persistence ----------------------------------- 1.71s",
                                "module-load : Persist modules via modules-load.d ------------------------ 1.68s",
                                "openvswitch : Flush Handlers -------------------------------------------- 1.68s",
                                "openvswitch : Waiting for openvswitch_db service to be ready ------------ 1.55s",
                                "openvswitch : Ensuring config directories exist ------------------------- 1.55s",
                                "Group hosts based on enabled services ----------------------------------- 1.50s",
                                "openvswitch : Copying over ovs-vsctl wrapper ---------------------------- 1.42s",
                                "Group hosts based on Kolla action --------------------------------------- 1.07s",
                                "openvswitch : Create /run/openvswitch directory on host ----------------- 0.73s",
                                "\u001b[32m2026-03-21 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask df62b1bb-5a17-4aaa-ad21-9610abb49f1f is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:48:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d647732c-4957-44b7-8e9a-a38699f22e53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "",
                                "\u001b[32m2026-03-21 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65bc1785-1a9b-4731-a631-cb5ede74952e is in state SUCCESS\u001b[0m",
                                "",
                                "PLAY [Prepare all k3s nodes] ***************************************************",
                                "",
                                "TASK [k3s_prereq : Validating arguments against arg spec 'main' - Prerequisites] ***",
                                "Saturday 21 March 2026  00:44:37 +0000 (0:00:00.261)       0:00:00.261 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Set same timezone on every Server] **************************",
                                "Saturday 21 March 2026  00:44:38 +0000 (0:00:00.552)       0:00:00.813 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Set SELinux to disabled state] ******************************",
                                "Saturday 21 March 2026  00:44:39 +0000 (0:00:00.759)       0:00:01.572 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Enable IPv4 forwarding] *************************************",
                                "Saturday 21 March 2026  00:44:39 +0000 (0:00:00.590)       0:00:02.163 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Enable IPv6 forwarding] *************************************",
                                "Saturday 21 March 2026  00:44:42 +0000 (0:00:02.271)       0:00:04.434 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Enable IPv6 router advertisements] **************************",
                                "Saturday 21 March 2026  00:44:43 +0000 (0:00:01.025)       0:00:05.460 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Add br_netfilter to /etc/modules-load.d/] *******************",
                                "Saturday 21 March 2026  00:44:44 +0000 (0:00:01.250)       0:00:06.711 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Load br_netfilter] ******************************************",
                                "Saturday 21 March 2026  00:44:45 +0000 (0:00:00.641)       0:00:07.352 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Set bridge-nf-call-iptables (just to be sure)] **************",
                                "Saturday 21 March 2026  00:44:45 +0000 (0:00:00.750)       0:00:08.103 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Add /usr/local/bin to sudo secure_path] *********************",
                                "Saturday 21 March 2026  00:44:46 +0000 (0:00:00.767)       0:00:08.870 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_download : Validating arguments against arg spec 'main' - Manage the downloading of K3S binaries] ***",
                                "Saturday 21 March 2026  00:44:47 +0000 (0:00:00.988)       0:00:09.858 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_download : Download k3s binary x64] **********************************",
                                "Saturday 21 March 2026  00:44:48 +0000 (0:00:01.196)       0:00:11.054 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_download : Download k3s binary arm64] ********************************",
                                "Saturday 21 March 2026  00:44:55 +0000 (0:00:06.510)       0:00:17.565 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_download : Download k3s binary armhf] ********************************",
                                "Saturday 21 March 2026  00:44:56 +0000 (0:00:01.176)       0:00:18.741 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_custom_registries : Validating arguments against arg spec 'main' - Configure the use of a custom container registry] ***",
                                "Saturday 21 March 2026  00:44:59 +0000 (0:00:02.999)       0:00:21.741 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_custom_registries : Create directory /etc/rancher/k3s] ***************",
                                "Saturday 21 March 2026  00:45:00 +0000 (0:00:01.399)       0:00:23.140 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_custom_registries : Insert registries into /etc/rancher/k3s/registries.yaml] ***",
                                "Saturday 21 March 2026  00:45:01 +0000 (0:00:01.028)       0:00:24.169 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_custom_registries : Remove /etc/rancher/k3s/registries.yaml when no registries configured] ***",
                                "Saturday 21 March 2026  00:45:03 +0000 (0:00:01.311)       0:00:25.481 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Deploy k3s master nodes] *************************************************",
                                "",
                                "TASK [k3s_server : Validating arguments against arg spec 'main' - Setup k3s servers] ***",
                                "Saturday 21 March 2026  00:45:04 +0000 (0:00:01.589)       0:00:27.070 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Stop k3s-init] **********************************************",
                                "Saturday 21 March 2026  00:45:06 +0000 (0:00:01.640)       0:00:28.711 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_server : Stop k3s] ***************************************************",
                                "Saturday 21 March 2026  00:45:08 +0000 (0:00:02.034)       0:00:30.746 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Clean previous runs of k3s-init] ****************************",
                                "Saturday 21 March 2026  00:45:09 +0000 (0:00:00.978)       0:00:31.724 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Deploy K3s http_proxy conf] *********************************",
                                "Saturday 21 March 2026  00:45:10 +0000 (0:00:01.457)       0:00:33.182 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create /etc/rancher/k3s directory] **************************",
                                "Saturday 21 March 2026  00:45:11 +0000 (0:00:00.695)       0:00:33.877 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create custom resolv.conf for k3s] **************************",
                                "Saturday 21 March 2026  00:45:12 +0000 (0:00:00.936)       0:00:34.813 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Deploy vip manifest] ****************************************",
                                "Saturday 21 March 2026  00:45:14 +0000 (0:00:01.635)       0:00:36.449 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/k3s_server/tasks/vip.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [k3s_server : Set _kube_vip_bgp_peers fact] *******************************",
                                "Saturday 21 March 2026  00:45:14 +0000 (0:00:00.646)       0:00:37.096 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create manifests directory on first master] *****************",
                                "Saturday 21 March 2026  00:45:16 +0000 (0:00:02.140)       0:00:39.236 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Download vip rbac manifest to first master] *****************",
                                "Saturday 21 March 2026  00:45:17 +0000 (0:00:00.599)       0:00:39.836 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Copy vip manifest to first master] **************************",
                                "Saturday 21 March 2026  00:45:18 +0000 (0:00:01.186)       0:00:41.022 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Deploy metallb manifest] ************************************",
                                "Saturday 21 March 2026  00:45:20 +0000 (0:00:01.319)       0:00:42.342 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Deploy kube-vip manifest] ***********************************",
                                "Saturday 21 March 2026  00:45:20 +0000 (0:00:00.638)       0:00:42.981 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Init cluster inside the transient k3s-init service] *********",
                                "Saturday 21 March 2026  00:45:21 +0000 (0:00:00.380)       0:00:43.361 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_server : Detect Kubernetes version for label compatibility] **********",
                                "Saturday 21 March 2026  00:45:24 +0000 (0:00:03.428)       0:00:46.791 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_server : Set node role label selector based on Kubernetes version] ***",
                                "Saturday 21 March 2026  00:45:27 +0000 (0:00:02.672)       0:00:49.464 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails)] ***",
                                "Saturday 21 March 2026  00:45:28 +0000 (0:00:00.869)       0:00:50.334 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Save logs of k3s-init.service] ******************************",
                                "Saturday 21 March 2026  00:46:11 +0000 (0:00:43.380)       0:01:33.714 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Kill the temporary service used for initialization] *********",
                                "Saturday 21 March 2026  00:46:11 +0000 (0:00:00.435)       0:01:34.149 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Copy K3s service file] **************************************",
                                "Saturday 21 March 2026  00:46:12 +0000 (0:00:00.845)       0:01:34.995 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Enable and check K3s service] *******************************",
                                "Saturday 21 March 2026  00:46:14 +0000 (0:00:01.401)       0:01:36.396 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Wait for node-token] ****************************************",
                                "Saturday 21 March 2026  00:46:39 +0000 (0:00:25.365)       0:02:01.762 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Register node-token file access mode] ***********************",
                                "Saturday 21 March 2026  00:46:40 +0000 (0:00:00.714)       0:02:02.477 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Change file access node-token] ******************************",
                                "Saturday 21 March 2026  00:46:42 +0000 (0:00:01.914)       0:02:04.391 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Read node-token from master] ********************************",
                                "Saturday 21 March 2026  00:46:42 +0000 (0:00:00.702)       0:02:05.094 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Store Master node-token] ************************************",
                                "Saturday 21 March 2026  00:46:43 +0000 (0:00:00.630)       0:02:05.725 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Restore node-token file access] *****************************",
                                "Saturday 21 March 2026  00:46:43 +0000 (0:00:00.541)       0:02:06.266 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create directory .kube] *************************************",
                                "Saturday 21 March 2026  00:46:44 +0000 (0:00:00.968)       0:02:07.235 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Copy config file to user home directory] ********************",
                                "Saturday 21 March 2026  00:46:45 +0000 (0:00:00.808)       0:02:08.043 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Configure kubectl cluster to https://192.168.16.8:6443] *****",
                                "Saturday 21 March 2026  00:46:46 +0000 (0:00:00.872)       0:02:08.915 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create kubectl symlink] *************************************",
                                "Saturday 21 March 2026  00:46:47 +0000 (0:00:00.827)       0:02:09.743 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create crictl symlink] **************************************",
                                "Saturday 21 March 2026  00:46:47 +0000 (0:00:00.415)       0:02:10.158 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Get contents of manifests folder] ***************************",
                                "Saturday 21 March 2026  00:46:48 +0000 (0:00:00.263)       0:02:10.422 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Get sub dirs of manifests folder] ***************************",
                                "Saturday 21 March 2026  00:46:48 +0000 (0:00:00.618)       0:02:11.040 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start] ***",
                                "Saturday 21 March 2026  00:46:49 +0000 (0:00:00.623)       0:02:11.664 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip-rbac.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m",
                                "",
                                "PLAY [Deploy k3s worker nodes] *************************************************",
                                "",
                                "TASK [k3s_agent : Validating arguments against arg spec 'main' - Setup k3s agents] ***",
                                "Saturday 21 March 2026  00:46:52 +0000 (0:00:02.785)       0:02:14.449 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Check if system is PXE-booted] *******************************",
                                "Saturday 21 March 2026  00:46:52 +0000 (0:00:00.311)       0:02:14.760 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Set fact for PXE-booted system] ******************************",
                                "Saturday 21 March 2026  00:46:53 +0000 (0:00:00.553)       0:02:15.314 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Include http_proxy configuration tasks] **********************",
                                "Saturday 21 March 2026  00:46:53 +0000 (0:00:00.463)       0:02:15.778 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/k3s_agent/tasks/http_proxy.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [k3s_agent : Create k3s-node.service.d directory] *************************",
                                "Saturday 21 March 2026  00:46:53 +0000 (0:00:00.449)       0:02:16.227 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Copy K3s http_proxy conf file] *******************************",
                                "Saturday 21 March 2026  00:46:54 +0000 (0:00:00.251)       0:02:16.479 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Deploy K3s http_proxy conf] **********************************",
                                "Saturday 21 March 2026  00:46:54 +0000 (0:00:00.386)       0:02:16.865 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Create /etc/rancher/k3s directory] ***************************",
                                "Saturday 21 March 2026  00:46:54 +0000 (0:00:00.294)       0:02:17.159 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Create custom resolv.conf for k3s] ***************************",
                                "Saturday 21 March 2026  00:46:55 +0000 (0:00:00.650)       0:02:17.810 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Configure the k3s service] ***********************************",
                                "Saturday 21 March 2026  00:46:56 +0000 (0:00:00.959)       0:02:18.769 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Manage k3s service] ******************************************",
                                "Saturday 21 March 2026  00:46:58 +0000 (0:00:01.682)       0:02:20.452 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Prepare kubeconfig file] *************************************************",
                                "",
                                "TASK [Get home directory of operator user] *************************************",
                                "Saturday 21 March 2026  00:47:08 +0000 (0:00:10.659)       0:02:31.112 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Create .kube directory] **************************************************",
                                "Saturday 21 March 2026  00:47:09 +0000 (0:00:00.785)       0:02:31.898 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Get kubeconfig file] *****************************************************",
                                "Saturday 21 March 2026  00:47:10 +0000 (0:00:00.410)       0:02:32.309 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [Write kubeconfig file] ***************************************************",
                                "Saturday 21 March 2026  00:47:10 +0000 (0:00:00.543)       0:02:32.852 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig] *********************************",
                                "Saturday 21 March 2026  00:47:11 +0000 (0:00:00.988)       0:02:33.841 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Make kubeconfig available for use inside the manager service] ************",
                                "Saturday 21 March 2026  00:47:12 +0000 (0:00:00.522)       0:02:34.363 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig inside the manager service] ******",
                                "Saturday 21 March 2026  00:47:13 +0000 (0:00:01.688)       0:02:36.052 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Set KUBECONFIG environment variable] *************************************",
                                "Saturday 21 March 2026  00:47:14 +0000 (0:00:00.897)       0:02:36.950 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Enable kubectl command line completion] **********************************",
                                "Saturday 21 March 2026  00:47:15 +0000 (0:00:00.394)       0:02:37.344 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Apply role kubectl] ******************************************************",
                                "",
                                "TASK [kubectl : Gather variables for each operating system] ********************",
                                "Saturday 21 March 2026  00:47:15 +0000 (0:00:00.395)       0:02:37.739 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Include distribution specific install tasks] *******************",
                                "Saturday 21 March 2026  00:47:15 +0000 (0:00:00.127)       0:02:37.867 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/kubectl/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [kubectl : Remove old architecture-dependent repository] ******************",
                                "Saturday 21 March 2026  00:47:15 +0000 (0:00:00.202)       0:02:38.070 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Install apt-transport-https package] ***************************",
                                "Saturday 21 March 2026  00:47:16 +0000 (0:00:01.053)       0:02:39.123 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Add repository gpg key] ****************************************",
                                "Saturday 21 March 2026  00:47:18 +0000 (0:00:01.300)       0:02:40.424 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Set permissions of gpg key] ************************************",
                                "Saturday 21 March 2026  00:47:19 +0000 (0:00:00.958)       0:02:41.383 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Add repository Debian] *****************************************",
                                "Saturday 21 March 2026  00:47:19 +0000 (0:00:00.503)       0:02:41.887 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Install required packages] *************************************",
                                "Saturday 21 March 2026  00:47:26 +0000 (0:00:06.754)       0:02:48.641 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Remove kubectl symlink] ****************************************",
                                "Saturday 21 March 2026  00:47:40 +0000 (0:00:14.199)       0:03:02.841 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Run post actions on master nodes] ****************************************",
                                "",
                                "TASK [k3s_server_post : Validating arguments against arg spec 'main' - Configure k3s cluster] ***",
                                "Saturday 21 March 2026  00:47:41 +0000 (0:00:00.525)       0:03:03.366 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Deploy calico] *****************************************",
                                "Saturday 21 March 2026  00:47:41 +0000 (0:00:00.502)       0:03:03.868 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Deploy cilium] *****************************************",
                                "Saturday 21 March 2026  00:47:41 +0000 (0:00:00.342)       0:03:04.211 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/k3s_server_post/tasks/cilium.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Create tmp directory on first master] ******************",
                                "Saturday 21 March 2026  00:47:42 +0000 (0:00:00.572)       0:03:04.783 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Wait for connectivity to kube VIP] *********************",
                                "Saturday 21 March 2026  00:47:43 +0000 (0:00:00.853)       0:03:05.637 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Fail if kube VIP not reachable] ************************",
                                "Saturday 21 March 2026  00:47:44 +0000 (0:00:00.938)       0:03:06.576 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Test for existing Cilium install] **********************",
                                "Saturday 21 March 2026  00:47:44 +0000 (0:00:00.282)       0:03:06.858 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Check Cilium version] **********************************",
                                "Saturday 21 March 2026  00:47:45 +0000 (0:00:01.059)       0:03:07.918 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Parse installed Cilium version] ************************",
                                "Saturday 21 March 2026  00:47:45 +0000 (0:00:00.101)       0:03:08.020 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Determine if Cilium needs update] **********************",
                                "Saturday 21 March 2026  00:47:45 +0000 (0:00:00.122)       0:03:08.142 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Log result] ********************************************",
                                "Saturday 21 March 2026  00:47:45 +0000 (0:00:00.105)       0:03:08.247 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Install Cilium] ****************************************",
                                "Saturday 21 March 2026  00:47:46 +0000 (0:00:00.112)       0:03:08.360 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Wait for Cilium resources] *****************************",
                                "Saturday 21 March 2026  00:47:50 +0000 (0:00:04.198)       0:03:12.558 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/cilium-operator)\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0 -> localhost]: Wait for Cilium resources (30 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=daemonset/cilium)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/hubble-relay)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/hubble-ui)\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Set _cilium_bgp_neighbors fact] ************************",
                                "Saturday 21 March 2026  00:48:33 +0000 (0:00:42.773)       0:03:55.332 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Copy BGP manifests to first master] ********************",
                                "Saturday 21 March 2026  00:48:34 +0000 (0:00:01.129)       0:03:56.462 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Apply BGP manifests] ***********************************",
                                "Saturday 21 March 2026  00:48:35 +0000 (0:00:01.569)       0:03:58.031 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Print error message if BGP manifests application fails] ***",
                                "Saturday 21 March 2026  00:48:36 +0000 (0:00:00.952)       0:03:58.983 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Test for BGP config resources] *************************",
                                "Saturday 21 March 2026  00:48:36 +0000 (0:00:00.114)       0:03:59.097 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=kubectl get CiliumBGPPeeringPolicy.cilium.io)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=kubectl get CiliumLoadBalancerIPPool.cilium.io)\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Deploy metallb pool] ***********************************",
                                "Saturday 21 March 2026  00:48:38 +0000 (0:00:01.792)       0:04:00.889 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Remove tmp directory used for manifests] ***************",
                                "Saturday 21 March 2026  00:48:38 +0000 (0:00:00.319)       0:04:01.209 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role k9s] **********************************************************",
                                "",
                                "TASK [k9s : Gather variables for each operating system] ************************",
                                "Saturday 21 March 2026  00:48:40 +0000 (0:00:01.128)       0:04:02.337 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [k9s : Include distribution specific install tasks] ***********************",
                                "Saturday 21 March 2026  00:48:40 +0000 (0:00:00.169)       0:04:02.507 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/k9s/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [k9s : Install k9s packages] **********************************************",
                                "Saturday 21 March 2026  00:48:40 +0000 (0:00:00.412)       0:04:02.920 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Manage labels, annotations, and taints on all k3s nodes] *****************",
                                "",
                                "TASK [Merge labels, annotations, and taints] ***********************************",
                                "Saturday 21 March 2026  00:48:46 +0000 (0:00:05.891)       0:04:08.812 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Manage labels] ***********************************************************",
                                "Saturday 21 March 2026  00:48:47 +0000 (0:00:00.606)       0:04:09.418 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m",
                                "",
                                "TASK [Manage annotations] ******************************************************",
                                "Saturday 21 March 2026  00:48:58 +0000 (0:00:11.474)       0:04:20.892 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Manage taints] ***********************************************************",
                                "Saturday 21 March 2026  00:48:59 +0000 (0:00:00.454)       0:04:21.347 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=50  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=28  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:48:59 +0000 (0:00:00.526)       0:04:21.874 ******** ",
                                "=============================================================================== ",
                                "k3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails) -- 43.38s",
                                "k3s_server_post : Wait for Cilium resources ---------------------------- 42.77s",
                                "k3s_server : Enable and check K3s service ------------------------------ 25.37s",
                                "kubectl : Install required packages ------------------------------------ 14.20s",
                                "Manage labels ---------------------------------------------------------- 11.47s",
                                "k3s_agent : Manage k3s service ----------------------------------------- 10.66s",
                                "kubectl : Add repository Debian ----------------------------------------- 6.75s",
                                "k3s_download : Download k3s binary x64 ---------------------------------- 6.51s",
                                "k9s : Install k9s packages ---------------------------------------------- 5.89s",
                                "k3s_server_post : Install Cilium ---------------------------------------- 4.20s",
                                "k3s_server : Init cluster inside the transient k3s-init service --------- 3.43s",
                                "k3s_download : Download k3s binary armhf -------------------------------- 3.00s",
                                "k3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start --- 2.79s",
                                "k3s_server : Detect Kubernetes version for label compatibility ---------- 2.67s",
                                "k3s_prereq : Enable IPv4 forwarding ------------------------------------- 2.27s",
                                "k3s_server : Set _kube_vip_bgp_peers fact ------------------------------- 2.14s",
                                "k3s_server : Stop k3s-init ---------------------------------------------- 2.03s",
                                "k3s_server : Register node-token file access mode ----------------------- 1.91s",
                                "k3s_server_post : Test for BGP config resources ------------------------- 1.79s",
                                "Make kubeconfig available for use inside the manager service ------------ 1.69s",
                                "\u001b[32m2026-03-21 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 591381dc-15c5-47b5-8a6a-1c0023f4941b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d647732c-4957-44b7-8e9a-a38699f22e53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 591381dc-15c5-47b5-8a6a-1c0023f4941b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d647732c-4957-44b7-8e9a-a38699f22e53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 591381dc-15c5-47b5-8a6a-1c0023f4941b is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d647732c-4957-44b7-8e9a-a38699f22e53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d647732c-4957-44b7-8e9a-a38699f22e53 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dbf2a30a-bdb2-4fea-8b1e-a0e8c3727cbc is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Copy kubeconfig to the configuration repository] *************************",
                                "",
                                "TASK [Get kubeconfig file] *****************************************************",
                                "Saturday 21 March 2026  00:49:02 +0000 (0:00:00.230)       0:00:00.230 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [Write kubeconfig file] ***************************************************",
                                "Saturday 21 March 2026  00:49:04 +0000 (0:00:01.458)       0:00:01.688 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig file] ****************************",
                                "Saturday 21 March 2026  00:49:06 +0000 (0:00:01.658)       0:00:03.347 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:49:06 +0000 (0:00:00.707)       0:00:04.055 ******** ",
                                "=============================================================================== ",
                                "Write kubeconfig file --------------------------------------------------- 1.66s",
                                "Get kubeconfig file ----------------------------------------------------- 1.46s",
                                "Change server address in the kubeconfig file ---------------------------- 0.71s",
                                "",
                                "",
                                "PLAY [Prepare kubeconfig file] *************************************************",
                                "",
                                "TASK [Get home directory of operator user] *************************************",
                                "Saturday 21 March 2026  00:49:02 +0000 (0:00:00.225)       0:00:00.225 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Create .kube directory] **************************************************",
                                "Saturday 21 March 2026  00:49:03 +0000 (0:00:00.764)       0:00:00.990 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Get kubeconfig file] *****************************************************",
                                "Saturday 21 March 2026  00:49:03 +0000 (0:00:00.495)       0:00:01.485 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [Write kubeconfig file] ***************************************************",
                                "Saturday 21 March 2026  00:49:04 +0000 (0:00:00.985)       0:00:02.471 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig] *********************************",
                                "Saturday 21 March 2026  00:49:05 +0000 (0:00:01.111)       0:00:03.582 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Make kubeconfig available for use inside the manager service] ************",
                                "Saturday 21 March 2026  00:49:06 +0000 (0:00:00.732)       0:00:04.314 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig inside the manager service] ******",
                                "Saturday 21 March 2026  00:49:08 +0000 (0:00:01.848)       0:00:06.163 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Set KUBECONFIG environment variable] *************************************",
                                "Saturday 21 March 2026  00:49:09 +0000 (0:00:01.482)       0:00:07.646 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Enable kubectl command line completion] **********************************",
                                "Saturday 21 March 2026  00:49:10 +0000 (0:00:00.518)       0:00:08.165 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:49:10 +0000 (0:00:00.528)       0:00:08.693 ******** ",
                                "=============================================================================== ",
                                "Make kubeconfig available for use inside the manager service ------------ 1.85s",
                                "Change server address in the kubeconfig inside the manager service ------ 1.48s",
                                "Write kubeconfig file --------------------------------------------------- 1.11s",
                                "Get kubeconfig file ----------------------------------------------------- 0.99s",
                                "Get home directory of operator user ------------------------------------- 0.76s",
                                "Change server address in the kubeconfig --------------------------------- 0.73s",
                                "Enable kubectl command line completion ---------------------------------- 0.53s",
                                "Set KUBECONFIG environment variable ------------------------------------- 0.52s",
                                "Create .kube directory -------------------------------------------------- 0.50s",
                                "",
                                "",
                                "PLAY [Set kolla_action_rabbitmq] ***********************************************",
                                "",
                                "TASK [Inform the user about the following task] ********************************",
                                "Saturday 21 March 2026  00:47:40 +0000 (0:00:00.096)       0:00:00.096 ******** ",
                                "\u001b[0;32mok: [localhost] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"The task 'Check RabbitMQ service' fails if the RabbitMQ service has not yet been deployed. This is fine.\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Check RabbitMQ service] **************************************************",
                                "Saturday 21 March 2026  00:47:40 +0000 (0:00:00.048)       0:00:00.145 ******** ",
                                "\u001b[0;31mfatal: [localhost]: FAILED! => {\"changed\": false, \"elapsed\": 2, \"msg\": \"Timeout when waiting for search string RabbitMQ Management in 192.168.16.9:15672\"}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "",
                                "TASK [Set kolla_action_rabbitmq = upgrade if RabbitMQ is already running] ******",
                                "Saturday 21 March 2026  00:47:45 +0000 (0:00:04.754)       0:00:04.899 ******** ",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "TASK [Set kolla_action_rabbitmq = kolla_action_ng] *****************************",
                                "Saturday 21 March 2026  00:47:45 +0000 (0:00:00.063)       0:00:04.963 ******** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  00:47:45 +0000 (0:00:00.234)       0:00:05.198 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  00:47:45 +0000 (0:00:00.343)       0:00:05.541 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_rabbitmq_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_rabbitmq_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_rabbitmq_True)\u001b[0m",
                                "",
                                "PLAY [Apply role rabbitmq] *****************************************************",
                                "",
                                "TASK [rabbitmq : include_tasks] ************************************************",
                                "Saturday 21 March 2026  00:47:46 +0000 (0:00:00.402)       0:00:05.944 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [rabbitmq : Get container facts] ******************************************",
                                "Saturday 21 March 2026  00:47:46 +0000 (0:00:00.563)       0:00:06.508 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Get current RabbitMQ version] *********************************",
                                "Saturday 21 March 2026  00:47:48 +0000 (0:00:01.788)       0:00:08.297 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Get new RabbitMQ version] *************************************",
                                "Saturday 21 March 2026  00:47:48 +0000 (0:00:00.399)       0:00:08.696 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Check if running RabbitMQ is at most one version behind] ******",
                                "Saturday 21 March 2026  00:47:49 +0000 (0:00:00.394)       0:00:09.090 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Catch when RabbitMQ is being downgraded] **********************",
                                "Saturday 21 March 2026  00:47:49 +0000 (0:00:00.375)       0:00:09.466 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : include_tasks] ************************************************",
                                "Saturday 21 March 2026  00:47:50 +0000 (0:00:00.475)       0:00:09.941 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/remove-ha-all-policy.yml for testbed-node-1, testbed-node-0, testbed-node-2\u001b[0m",
                                "",
                                "TASK [rabbitmq : Get container facts] ******************************************",
                                "Saturday 21 March 2026  00:47:51 +0000 (0:00:01.233)       0:00:11.175 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : List RabbitMQ policies] ***************************************",
                                "Saturday 21 March 2026  00:47:52 +0000 (0:00:01.319)       0:00:12.495 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Remove ha-all policy from RabbitMQ] ***************************",
                                "Saturday 21 March 2026  00:47:54 +0000 (0:00:01.861)       0:00:14.356 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Ensuring config directories exist] ****************************",
                                "Saturday 21 March 2026  00:47:55 +0000 (0:00:00.652)       0:00:15.009 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over config.json files for services] ******************",
                                "Saturday 21 March 2026  00:47:56 +0000 (0:00:01.202)       0:00:16.212 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over rabbitmq-env.conf] *******************************",
                                "Saturday 21 March 2026  00:47:57 +0000 (0:00:01.557)       0:00:17.770 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over rabbitmq.conf] ***********************************",
                                "Saturday 21 March 2026  00:47:59 +0000 (0:00:01.801)       0:00:19.572 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over erl_inetrc] **************************************",
                                "Saturday 21 March 2026  00:48:02 +0000 (0:00:02.319)       0:00:21.892 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over advanced.config] *********************************",
                                "Saturday 21 March 2026  00:48:03 +0000 (0:00:01.216)       0:00:23.108 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over definitions.json] ********************************",
                                "Saturday 21 March 2026  00:48:04 +0000 (0:00:01.681)       0:00:24.790 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over enabled_plugins] *********************************",
                                "Saturday 21 March 2026  00:48:06 +0000 (0:00:01.400)       0:00:26.190 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : include_tasks] ************************************************",
                                "Saturday 21 March 2026  00:48:08 +0000 (0:00:01.813)       0:00:28.004 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Check rabbitmq containers] ************************************",
                                "Saturday 21 March 2026  00:48:08 +0000 (0:00:00.397)       0:00:28.402 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [rabbitmq : Creating rabbitmq volume] *************************************",
                                "Saturday 21 March 2026  00:48:09 +0000 (0:00:01.001)       0:00:29.403 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Running RabbitMQ bootstrap container] *************************",
                                "Saturday 21 March 2026  00:48:10 +0000 (0:00:00.821)       0:00:30.224 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [rabbitmq : Restart rabbitmq container] ************************",
                                "Saturday 21 March 2026  00:48:16 +0000 (0:00:06.209)       0:00:36.434 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Restart rabbitmq services] ***********************************************",
                                "",
                                "TASK [rabbitmq : Get info on RabbitMQ container] *******************************",
                                "Saturday 21 March 2026  00:48:16 +0000 (0:00:00.349)       0:00:36.784 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************",
                                "Saturday 21 March 2026  00:48:17 +0000 (0:00:00.773)       0:00:37.557 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Restart rabbitmq container] ***********************************",
                                "Saturday 21 March 2026  00:48:18 +0000 (0:00:00.600)       0:00:38.158 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Waiting for rabbitmq to start] ********************************",
                                "Saturday 21 March 2026  00:48:20 +0000 (0:00:01.989)       0:00:40.148 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Restart rabbitmq services] ***********************************************",
                                "",
                                "TASK [rabbitmq : Get info on RabbitMQ container] *******************************",
                                "Saturday 21 March 2026  00:49:13 +0000 (0:00:53.364)       0:01:33.512 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************",
                                "Saturday 21 March 2026  00:49:14 +0000 (0:00:00.537)       0:01:34.050 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Restart rabbitmq container] ***********************************",
                                "Saturday 21 March 2026  00:49:14 +0000 (0:00:00.225)       0:01:34.276 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Waiting for rabbitmq to start] ********************************",
                                "Saturday 21 March 2026  00:49:16 +0000 (0:00:02.332)       0:01:36.609 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Restart rabbitmq services] ***********************************************",
                                "",
                                "TASK [rabbitmq : Get info on RabbitMQ container] *******************************",
                                "Saturday 21 March 2026  00:49:29 +0000 (0:00:13.207)       0:01:49.817 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************",
                                "Saturday 21 March 2026  00:49:30 +0000 (0:00:00.589)       0:01:50.406 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Restart rabbitmq container] ***********************************",
                                "Saturday 21 March 2026  00:49:30 +0000 (0:00:00.217)       0:01:50.624 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Waiting for rabbitmq to start] ********************************",
                                "Saturday 21 March 2026  00:49:37 +0000 (0:00:06.464)       0:01:57.088 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply rabbitmq post-configuration] ***************************************",
                                "",
                                "TASK [Include rabbitmq post-deploy.yml] ****************************************",
                                "Saturday 21 March 2026  00:49:48 +0000 (0:00:10.785)       0:02:07.874 ******** ",
                                "\u001b[0;36mincluded: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [rabbitmq : Enable all stable feature flags] ******************************",
                                "Saturday 21 March 2026  00:49:48 +0000 (0:00:00.663)       0:02:08.537 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m",
                                "\u001b[1;35menable_outward_rabbitmq_True\u001b[0m",
                                "",
                                "PLAY [Apply role rabbitmq (outward)] *******************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m",
                                "\u001b[1;35moutward_rabbitmq_restart\u001b[0m",
                                "",
                                "PLAY [Restart rabbitmq (outward) services] *************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Apply rabbitmq (outward) post-configuration] *****************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mlocalhost\u001b[0m                  : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:49:51 +0000 (0:00:02.529)       0:02:11.066 ******** ",
                                "=============================================================================== ",
                                "rabbitmq : Waiting for rabbitmq to start ------------------------------- 77.36s",
                                "rabbitmq : Restart rabbitmq container ---------------------------------- 10.79s",
                                "rabbitmq : Running RabbitMQ bootstrap container ------------------------- 6.21s",
                                "Check RabbitMQ service -------------------------------------------------- 4.75s",
                                "rabbitmq : Enable all stable feature flags ------------------------------ 2.53s",
                                "rabbitmq : Copying over rabbitmq.conf ----------------------------------- 2.32s",
                                "rabbitmq : Get info on RabbitMQ container ------------------------------- 1.90s",
                                "rabbitmq : List RabbitMQ policies --------------------------------------- 1.86s",
                                "rabbitmq : Copying over enabled_plugins --------------------------------- 1.81s",
                                "rabbitmq : Copying over rabbitmq-env.conf ------------------------------- 1.80s",
                                "rabbitmq : Get container facts ------------------------------------------ 1.79s",
                                "rabbitmq : Copying over advanced.config --------------------------------- 1.68s",
                                "rabbitmq : Copying over config.json files for services ------------------ 1.56s",
                                "rabbitmq : Copying over definitions.json -------------------------------- 1.40s",
                                "rabbitmq : Get container facts ------------------------------------------ 1.32s",
                                "rabbitmq : include_tasks ------------------------------------------------ 1.23s",
                                "rabbitmq : Copying over erl_inetrc -------------------------------------- 1.22s",
                                "rabbitmq : Ensuring config directories exist ---------------------------- 1.20s",
                                "rabbitmq : Put RabbitMQ node into maintenance mode ---------------------- 1.04s",
                                "rabbitmq : Check rabbitmq containers ------------------------------------ 1.00s",
                                "\u001b[32m2026-03-21 00:49:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:49:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 69c2b83b-6192-4a92-bea8-8a6b37255959 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  00:48:28 +0000 (0:00:00.161)       0:00:00.161 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  00:48:29 +0000 (0:00:00.799)       0:00:00.961 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_ovn_True)\u001b[0m",
                                "",
                                "PLAY [Apply role ovn-controller] ***********************************************",
                                "",
                                "TASK [ovn-controller : include_tasks] ******************************************",
                                "Saturday 21 March 2026  00:48:30 +0000 (0:00:01.358)       0:00:02.320 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-controller/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ovn-controller : Ensuring config directories exist] **********************",
                                "Saturday 21 March 2026  00:48:31 +0000 (0:00:01.116)       0:00:03.436 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Copying over config.json files for services] ************",
                                "Saturday 21 March 2026  00:48:33 +0000 (0:00:01.531)       0:00:04.967 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Ensuring systemd override directory exists] *************",
                                "Saturday 21 March 2026  00:48:35 +0000 (0:00:01.747)       0:00:06.715 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Copying over systemd override] **************************",
                                "Saturday 21 March 2026  00:48:37 +0000 (0:00:01.955)       0:00:08.670 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Check ovn-controller containers] ************************",
                                "Saturday 21 March 2026  00:48:38 +0000 (0:00:01.573)       0:00:10.244 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Create br-int bridge on OpenvSwitch] ********************",
                                "Saturday 21 March 2026  00:48:40 +0000 (0:00:02.032)       0:00:12.276 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-controller : Configure OVN in OVSDB] *********************************",
                                "Saturday 21 March 2026  00:48:43 +0000 (0:00:02.796)       0:00:15.073 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.13'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.12'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.14'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.10'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.15'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.11'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:89:18:56', 'state': 'present'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:29:4a:9b', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:2f:fa:44', 'state': 'present'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:52:c1:40', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:71:3a:c3', 'state': 'present'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:33:12:50', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Saturday 21 March 2026  00:49:02 +0000 (0:00:18.425)       0:00:33.499 ******** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Saturday 21 March 2026  00:49:02 +0000 (0:00:00.090)       0:00:33.589 ******** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Saturday 21 March 2026  00:49:02 +0000 (0:00:00.067)       0:00:33.656 ******** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Saturday 21 March 2026  00:49:02 +0000 (0:00:00.064)       0:00:33.721 ******** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Saturday 21 March 2026  00:49:02 +0000 (0:00:00.058)       0:00:33.780 ******** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Saturday 21 March 2026  00:49:02 +0000 (0:00:00.071)       0:00:33.851 ******** ",
                                "",
                                "RUNNING HANDLER [ovn-controller : Reload systemd config] ***********************",
                                "Saturday 21 March 2026  00:49:02 +0000 (0:00:00.065)       0:00:33.917 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-controller : Restart ovn-controller container] ************",
                                "Saturday 21 March 2026  00:49:04 +0000 (0:00:01.643)       0:00:35.561 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY [Apply role ovn-db] *******************************************************",
                                "",
                                "TASK [ovn-db : include_tasks] **************************************************",
                                "Saturday 21 March 2026  00:49:31 +0000 (0:00:27.405)       0:01:02.967 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ovn-db : include_tasks] **************************************************",
                                "Saturday 21 March 2026  00:49:31 +0000 (0:00:00.446)       0:01:03.413 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/lookup_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ovn-db : Checking for any existing OVN DB container volumes] *************",
                                "Saturday 21 March 2026  00:49:32 +0000 (0:00:00.661)       0:01:04.075 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN NB volume availability] ***************",
                                "Saturday 21 March 2026  00:49:33 +0000 (0:00:00.999)       0:01:05.075 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN SB volume availability] ***************",
                                "Saturday 21 March 2026  00:49:33 +0000 (0:00:00.331)       0:01:05.407 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Establish whether the OVN NB cluster has already existed] *******",
                                "Saturday 21 March 2026  00:49:34 +0000 (0:00:00.490)       0:01:05.897 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Establish whether the OVN SB cluster has already existed] *******",
                                "Saturday 21 March 2026  00:49:34 +0000 (0:00:00.310)       0:01:06.208 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check if running on all OVN NB DB hosts] ************************",
                                "Saturday 21 March 2026  00:49:35 +0000 (0:00:00.292)       0:01:06.501 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check OVN NB service port liveness] *****************************",
                                "Saturday 21 March 2026  00:49:35 +0000 (0:00:00.289)       0:01:06.790 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN NB service port liveness] *************",
                                "Saturday 21 March 2026  00:49:35 +0000 (0:00:00.296)       0:01:07.086 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN NB database information] ********************************",
                                "Saturday 21 March 2026  00:49:36 +0000 (0:00:00.486)       0:01:07.573 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN NB leader/follower role] **************",
                                "Saturday 21 March 2026  00:49:36 +0000 (0:00:00.284)       0:01:07.857 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Fail on existing OVN NB cluster with no leader] *****************",
                                "Saturday 21 March 2026  00:49:36 +0000 (0:00:00.268)       0:01:08.126 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check if running on all OVN SB DB hosts] ************************",
                                "Saturday 21 March 2026  00:49:36 +0000 (0:00:00.270)       0:01:08.397 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check OVN SB service port liveness] *****************************",
                                "Saturday 21 March 2026  00:49:37 +0000 (0:00:00.427)       0:01:08.824 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN SB service port liveness] *************",
                                "Saturday 21 March 2026  00:49:37 +0000 (0:00:00.253)       0:01:09.078 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN SB database information] ********************************",
                                "Saturday 21 March 2026  00:49:37 +0000 (0:00:00.286)       0:01:09.364 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN SB leader/follower role] **************",
                                "Saturday 21 March 2026  00:49:38 +0000 (0:00:00.254)       0:01:09.618 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Fail on existing OVN SB cluster with no leader] *****************",
                                "Saturday 21 March 2026  00:49:38 +0000 (0:00:00.392)       0:01:10.011 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : include_tasks] **************************************************",
                                "Saturday 21 March 2026  00:49:38 +0000 (0:00:00.248)       0:01:10.259 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/bootstrap-initial.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ovn-db : Set bootstrap args fact for NB (new cluster)] *******************",
                                "Saturday 21 March 2026  00:49:39 +0000 (0:00:00.503)       0:01:10.763 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Set bootstrap args fact for SB (new cluster)] *******************",
                                "Saturday 21 March 2026  00:49:39 +0000 (0:00:00.630)       0:01:11.394 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check NB cluster status] ****************************************",
                                "Saturday 21 March 2026  00:49:40 +0000 (0:00:00.419)       0:01:11.814 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check SB cluster status] ****************************************",
                                "Saturday 21 March 2026  00:49:40 +0000 (0:00:00.341)       0:01:12.155 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Remove an old node with the same ip address as the new node in NB DB] ***",
                                "Saturday 21 March 2026  00:49:41 +0000 (0:00:00.407)       0:01:12.563 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Remove an old node with the same ip address as the new node in SB DB] ***",
                                "Saturday 21 March 2026  00:49:41 +0000 (0:00:00.483)       0:01:13.046 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Set bootstrap args fact for NB (new member)] ********************",
                                "Saturday 21 March 2026  00:49:41 +0000 (0:00:00.347)       0:01:13.393 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Set bootstrap args fact for SB (new member)] ********************",
                                "Saturday 21 March 2026  00:49:42 +0000 (0:00:00.269)       0:01:13.663 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Ensuring config directories exist] ******************************",
                                "Saturday 21 March 2026  00:49:42 +0000 (0:00:00.285)       0:01:13.948 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-db : Copying over config.json files for services] ********************",
                                "Saturday 21 March 2026  00:49:44 +0000 (0:00:01.587)       0:01:15.535 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-db : Check ovn containers] *******************************************",
                                "Saturday 21 March 2026  00:49:48 +0000 (0:00:04.183)       0:01:19.719 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Saturday 21 March 2026  00:49:50 +0000 (0:00:02.413)       0:01:22.133 ******** ",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Saturday 21 March 2026  00:49:50 +0000 (0:00:00.122)       0:01:22.255 ******** ",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Saturday 21 March 2026  00:49:50 +0000 (0:00:00.120)       0:01:22.376 ******** ",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-nb-db container] *************************",
                                "Saturday 21 March 2026  00:49:51 +0000 (0:00:00.134)       0:01:22.511 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-sb-db container] *************************",
                                "Saturday 21 March 2026  00:49:53 +0000 (0:00:02.712)       0:01:25.224 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-northd container] ************************",
                                "Saturday 21 March 2026  00:50:01 +0000 (0:00:07.696)       0:01:32.920 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for leader election] ***************************************",
                                "Saturday 21 March 2026  00:50:08 +0000 (0:00:07.540)       0:01:40.461 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN_Northbound cluster leader] ******************************",
                                "Saturday 21 March 2026  00:50:09 +0000 (0:00:00.109)       0:01:40.571 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Configure OVN NB connection settings] ***************************",
                                "Saturday 21 March 2026  00:50:11 +0000 (0:00:01.952)       0:01:42.523 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN_Southbound cluster leader] ******************************",
                                "Saturday 21 March 2026  00:50:11 +0000 (0:00:00.745)       0:01:43.269 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Configure OVN SB connection settings] ***************************",
                                "Saturday 21 March 2026  00:50:12 +0000 (0:00:01.058)       0:01:44.328 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-nb-db] *********************************************",
                                "Saturday 21 March 2026  00:50:13 +0000 (0:00:00.668)       0:01:44.996 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-sb-db] *********************************************",
                                "Saturday 21 March 2026  00:50:14 +0000 (0:00:00.991)       0:01:45.988 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Unset bootstrap args fact] **************************************",
                                "Saturday 21 March 2026  00:50:15 +0000 (0:00:00.998)       0:01:46.986 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Ensuring config directories exist] ******************************",
                                "Saturday 21 March 2026  00:50:16 +0000 (0:00:00.531)       0:01:47.518 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-db : Copying over config.json files for services] ********************",
                                "Saturday 21 March 2026  00:50:17 +0000 (0:00:01.500)       0:01:49.018 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-db : Check ovn containers] *******************************************",
                                "Saturday 21 March 2026  00:50:21 +0000 (0:00:03.571)       0:01:52.590 ******** ",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Saturday 21 March 2026  00:50:24 +0000 (0:00:02.940)       0:01:55.531 ******** ",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Saturday 21 March 2026  00:50:24 +0000 (0:00:00.066)       0:01:55.597 ******** ",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Saturday 21 March 2026  00:50:24 +0000 (0:00:00.065)       0:01:55.663 ******** ",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-nb-db container] *************************",
                                "Saturday 21 March 2026  00:50:24 +0000 (0:00:00.362)       0:01:56.025 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-sb-db container] *************************",
                                "Saturday 21 March 2026  00:50:30 +0000 (0:00:06.292)       0:02:02.318 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-northd container] ************************",
                                "Saturday 21 March 2026  00:50:37 +0000 (0:00:06.190)       0:02:08.509 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for leader election] ***************************************",
                                "Saturday 21 March 2026  00:50:43 +0000 (0:00:06.125)       0:02:14.634 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN_Northbound cluster leader] ******************************",
                                "Saturday 21 March 2026  00:50:43 +0000 (0:00:00.114)       0:02:14.749 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Configure OVN NB connection settings] ***************************",
                                "Saturday 21 March 2026  00:50:44 +0000 (0:00:00.757)       0:02:15.507 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN_Southbound cluster leader] ******************************",
                                "Saturday 21 March 2026  00:50:44 +0000 (0:00:00.703)       0:02:16.211 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Configure OVN SB connection settings] ***************************",
                                "Saturday 21 March 2026  00:50:45 +0000 (0:00:00.771)       0:02:16.982 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-nb-db] *********************************************",
                                "Saturday 21 March 2026  00:50:46 +0000 (0:00:00.705)       0:02:17.687 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-sb-db] *********************************************",
                                "Saturday 21 March 2026  00:50:47 +0000 (0:00:00.840)       0:02:18.528 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=44  \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=43  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=22  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=43  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=22  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:50:48 +0000 (0:00:01.190)       0:02:19.719 ******** ",
                                "=============================================================================== ",
                                "ovn-controller : Restart ovn-controller container ---------------------- 27.41s",
                                "ovn-controller : Configure OVN in OVSDB -------------------------------- 18.43s",
                                "ovn-db : Restart ovn-sb-db container ----------------------------------- 13.89s",
                                "ovn-db : Restart ovn-northd container ---------------------------------- 13.67s",
                                "ovn-db : Restart ovn-nb-db container ------------------------------------ 9.01s",
                                "ovn-db : Copying over config.json files for services -------------------- 4.18s",
                                "ovn-db : Copying over config.json files for services -------------------- 3.57s",
                                "ovn-db : Check ovn containers ------------------------------------------- 2.94s",
                                "ovn-controller : Create br-int bridge on OpenvSwitch -------------------- 2.80s",
                                "ovn-db : Check ovn containers ------------------------------------------- 2.41s",
                                "ovn-controller : Check ovn-controller containers ------------------------ 2.03s",
                                "ovn-controller : Ensuring systemd override directory exists ------------- 1.95s",
                                "ovn-db : Get OVN_Northbound cluster leader ------------------------------ 1.95s",
                                "ovn-controller : Copying over config.json files for services ------------ 1.75s",
                                "ovn-controller : Reload systemd config ---------------------------------- 1.64s",
                                "ovn-db : Ensuring config directories exist ------------------------------ 1.59s",
                                "ovn-controller : Copying over systemd override -------------------------- 1.57s",
                                "ovn-controller : Ensuring config directories exist ---------------------- 1.53s",
                                "ovn-db : Ensuring config directories exist ------------------------------ 1.50s",
                                "Group hosts based on enabled services ----------------------------------- 1.36s",
                                "\u001b[32m2026-03-21 00:50:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:50:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:51:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:52:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 617b59f7-a1d9-4489-96b3-407edd7963ed is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  00:47:19 +0000 (0:00:00.553)       0:00:00.553 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  00:47:19 +0000 (0:00:00.493)       0:00:01.046 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_loadbalancer_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_loadbalancer_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_loadbalancer_True)\u001b[0m",
                                "",
                                "PLAY [Apply role loadbalancer] *************************************************",
                                "",
                                "TASK [loadbalancer : include_tasks] ********************************************",
                                "Saturday 21 March 2026  00:47:20 +0000 (0:00:00.481)       0:00:01.528 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/loadbalancer/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [loadbalancer : Check IPv6 support] ***************************************",
                                "Saturday 21 March 2026  00:47:21 +0000 (0:00:01.501)       0:00:03.029 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Setting sysctl values] ***************************************************",
                                "Saturday 21 March 2026  00:47:23 +0000 (0:00:01.593)       0:00:04.623 ******** ",
                                "\u001b[0;36mincluded: sysctl for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [sysctl : Check IPv6 support] *********************************************",
                                "Saturday 21 March 2026  00:47:24 +0000 (0:00:00.707)       0:00:05.330 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [sysctl : Setting sysctl values] ******************************************",
                                "Saturday 21 March 2026  00:47:25 +0000 (0:00:01.345)       0:00:06.676 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m",
                                "",
                                "TASK [module-load : Load modules] **********************************************",
                                "Saturday 21 March 2026  00:47:28 +0000 (0:00:02.594)       0:00:09.271 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=ip_vs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=ip_vs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=ip_vs)\u001b[0m",
                                "",
                                "TASK [module-load : Persist modules via modules-load.d] ************************",
                                "Saturday 21 March 2026  00:47:29 +0000 (0:00:00.969)       0:00:10.240 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=ip_vs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=ip_vs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=ip_vs)\u001b[0m",
                                "",
                                "TASK [module-load : Drop module persistence] ***********************************",
                                "Saturday 21 March 2026  00:47:30 +0000 (0:00:01.394)       0:00:11.635 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=ip_vs) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=ip_vs) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=ip_vs) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Ensuring config directories exist] ************************",
                                "Saturday 21 March 2026  00:47:31 +0000 (0:00:00.820)       0:00:12.456 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [loadbalancer : Ensuring haproxy service config subdir exists] ************",
                                "Saturday 21 March 2026  00:47:32 +0000 (0:00:01.664)       0:00:14.120 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Ensuring proxysql service config subdirectories exist] ****",
                                "Saturday 21 March 2026  00:47:34 +0000 (0:00:01.526)       0:00:15.647 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=users)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=users)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=users)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=rules)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=rules)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=rules)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Ensuring keepalived checks subdir exists] *****************",
                                "Saturday 21 March 2026  00:47:36 +0000 (0:00:02.289)       0:00:17.937 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Remove mariadb.cfg if proxysql enabled] *******************",
                                "Saturday 21 March 2026  00:47:38 +0000 (0:00:01.409)       0:00:19.346 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Removing checks for services which are disabled] **********",
                                "Saturday 21 March 2026  00:47:40 +0000 (0:00:02.608)       0:00:21.955 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__dbae62d55c991b3958350c8dc51f9c592d9f63fa', '__omit_place_holder__dbae62d55c991b3958350c8dc51f9c592d9f63fa'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__dbae62d55c991b3958350c8dc51f9c592d9f63fa', '__omit_place_holder__dbae62d55c991b3958350c8dc51f9c592d9f63fa'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__dbae62d55c991b3958350c8dc51f9c592d9f63fa', '__omit_place_holder__dbae62d55c991b3958350c8dc51f9c592d9f63fa'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying checks for services which are enabled] ************",
                                "Saturday 21 March 2026  00:47:43 +0000 (0:00:02.608)       0:00:24.564 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__dbae62d55c991b3958350c8dc51f9c592d9f63fa', '__omit_place_holder__dbae62d55c991b3958350c8dc51f9c592d9f63fa'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__dbae62d55c991b3958350c8dc51f9c592d9f63fa', '__omit_place_holder__dbae62d55c991b3958350c8dc51f9c592d9f63fa'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__dbae62d55c991b3958350c8dc51f9c592d9f63fa', '__omit_place_holder__dbae62d55c991b3958350c8dc51f9c592d9f63fa'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over config.json files for services] **************",
                                "Saturday 21 March 2026  00:47:46 +0000 (0:00:03.184)       0:00:27.748 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy.cfg] *********************************",
                                "Saturday 21 March 2026  00:47:49 +0000 (0:00:02.868)       0:00:30.616 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over proxysql config] *****************************",
                                "Saturday 21 March 2026  00:47:51 +0000 (0:00:02.358)       0:00:32.975 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy single external frontend config] *****",
                                "Saturday 21 March 2026  00:47:57 +0000 (0:00:05.487)       0:00:38.463 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over custom haproxy services configuration] *******",
                                "Saturday 21 March 2026  00:47:58 +0000 (0:00:00.939)       0:00:39.402 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over keepalived.conf] *****************************",
                                "Saturday 21 March 2026  00:48:00 +0000 (0:00:02.389)       0:00:41.792 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy.pem] *********************************",
                                "Saturday 21 March 2026  00:48:02 +0000 (0:00:01.882)       0:00:43.675 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=haproxy.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=haproxy.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=haproxy.pem)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy-internal.pem] ************************",
                                "Saturday 21 March 2026  00:48:04 +0000 (0:00:01.630)       0:00:45.305 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=haproxy-internal.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=haproxy-internal.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=haproxy-internal.pem)\u001b[0m",
                                "",
                                "TASK [loadbalancer : include_tasks] ********************************************",
                                "Saturday 21 March 2026  00:48:05 +0000 (0:00:01.550)       0:00:46.856 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/loadbalancer/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : loadbalancer | Copying over extra CA certificates] ***",
                                "Saturday 21 March 2026  00:48:06 +0000 (0:00:00.834)       0:00:47.690 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : loadbalancer | Copying over backend internal TLS certificate] ***",
                                "Saturday 21 March 2026  00:48:09 +0000 (0:00:03.340)       0:00:51.031 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : loadbalancer | Copying over backend internal TLS key] ***",
                                "Saturday 21 March 2026  00:48:10 +0000 (0:00:00.531)       0:00:51.563 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over extra CA certificates] ********",
                                "Saturday 21 March 2026  00:48:11 +0000 (0:00:01.446)       0:00:53.010 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] ***",
                                "Saturday 21 March 2026  00:48:12 +0000 (0:00:00.653)       0:00:53.663 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over backend internal TLS key] *****",
                                "Saturday 21 March 2026  00:48:13 +0000 (0:00:00.593)       0:00:54.257 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : proxysql | Copying over extra CA certificates] *******",
                                "Saturday 21 March 2026  00:48:14 +0000 (0:00:01.238)       0:00:55.496 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : proxysql | Copying over backend internal TLS certificate] ***",
                                "Saturday 21 March 2026  00:48:14 +0000 (0:00:00.637)       0:00:56.134 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : proxysql | Copying over backend internal TLS key] ****",
                                "Saturday 21 March 2026  00:48:15 +0000 (0:00:00.599)       0:00:56.734 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy start script] ************************",
                                "Saturday 21 March 2026  00:48:16 +0000 (0:00:01.387)       0:00:58.122 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over proxysql start script] ***********************",
                                "Saturday 21 March 2026  00:48:19 +0000 (0:00:02.494)       0:01:00.617 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying files for haproxy-ssh] ****************************",
                                "Saturday 21 March 2026  00:48:20 +0000 (0:00:01.387)       0:01:02.004 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Check loadbalancer containers] ****************************",
                                "Saturday 21 March 2026  00:48:21 +0000 (0:00:00.861)       0:01:02.866 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [include_role : aodh] *****************************************************",
                                "Saturday 21 March 2026  00:48:23 +0000 (0:00:02.229)       0:01:05.095 ******** ",
                                "\u001b[0;36mincluded: aodh for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over aodh haproxy config] ***********************",
                                "Saturday 21 March 2026  00:48:24 +0000 (0:00:00.566)       0:01:05.662 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for aodh when using single external frontend] ***",
                                "Saturday 21 March 2026  00:48:27 +0000 (0:00:03.420)       0:01:09.082 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for aodh] **************************",
                                "Saturday 21 March 2026  00:48:28 +0000 (0:00:00.714)       0:01:09.797 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over aodh ProxySQL users config] ***************",
                                "Saturday 21 March 2026  00:48:29 +0000 (0:00:00.802)       0:01:10.600 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over aodh ProxySQL rules config] ***************",
                                "Saturday 21 March 2026  00:48:31 +0000 (0:00:01.829)       0:01:12.430 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : barbican] *************************************************",
                                "Saturday 21 March 2026  00:48:33 +0000 (0:00:02.207)       0:01:14.637 ******** ",
                                "\u001b[0;36mincluded: barbican for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over barbican haproxy config] *******************",
                                "Saturday 21 March 2026  00:48:34 +0000 (0:00:00.675)       0:01:15.313 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for barbican when using single external frontend] ***",
                                "Saturday 21 March 2026  00:48:38 +0000 (0:00:04.283)       0:01:19.597 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for barbican] **********************",
                                "Saturday 21 March 2026  00:48:39 +0000 (0:00:00.832)       0:01:20.429 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over barbican ProxySQL users config] ***********",
                                "Saturday 21 March 2026  00:48:40 +0000 (0:00:00.990)       0:01:21.419 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over barbican ProxySQL rules config] ***********",
                                "Saturday 21 March 2026  00:48:41 +0000 (0:00:01.311)       0:01:22.731 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : blazar] ***************************************************",
                                "Saturday 21 March 2026  00:48:43 +0000 (0:00:01.975)       0:01:24.707 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : ceph-rgw] *************************************************",
                                "Saturday 21 March 2026  00:48:43 +0000 (0:00:00.251)       0:01:24.959 ******** ",
                                "\u001b[0;36mincluded: ceph-rgw for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over ceph-rgw haproxy config] *******************",
                                "Saturday 21 March 2026  00:48:46 +0000 (0:00:02.307)       0:01:27.267 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for ceph-rgw when using single external frontend] ***",
                                "Saturday 21 March 2026  00:48:49 +0000 (0:00:02.946)       0:01:30.213 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for ceph-rgw] **********************",
                                "Saturday 21 March 2026  00:48:50 +0000 (0:00:01.939)       0:01:32.153 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over ceph-rgw ProxySQL users config] ***********",
                                "Saturday 21 March 2026  00:48:53 +0000 (0:00:02.249)       0:01:34.402 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over ceph-rgw ProxySQL rules config] ***********",
                                "Saturday 21 March 2026  00:48:53 +0000 (0:00:00.430)       0:01:34.832 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : cinder] ***************************************************",
                                "Saturday 21 March 2026  00:48:54 +0000 (0:00:00.958)       0:01:35.791 ******** ",
                                "\u001b[0;36mincluded: cinder for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over cinder haproxy config] *********************",
                                "Saturday 21 March 2026  00:48:55 +0000 (0:00:00.981)       0:01:36.773 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for cinder when using single external frontend] ***",
                                "Saturday 21 March 2026  00:48:59 +0000 (0:00:03.753)       0:01:40.526 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for cinder] ************************",
                                "Saturday 21 March 2026  00:48:59 +0000 (0:00:00.621)       0:01:41.148 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over cinder ProxySQL users config] *************",
                                "Saturday 21 March 2026  00:49:00 +0000 (0:00:01.009)       0:01:42.157 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over cinder ProxySQL rules config] *************",
                                "Saturday 21 March 2026  00:49:02 +0000 (0:00:01.053)       0:01:43.211 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [include_role : cloudkitty] ***********************************************",
                                "Saturday 21 March 2026  00:49:03 +0000 (0:00:01.647)       0:01:44.858 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : cyborg] ***************************************************",
                                "Saturday 21 March 2026  00:49:03 +0000 (0:00:00.238)       0:01:45.097 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : designate] ************************************************",
                                "Saturday 21 March 2026  00:49:04 +0000 (0:00:00.257)       0:01:45.354 ******** ",
                                "\u001b[0;36mincluded: designate for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over designate haproxy config] ******************",
                                "Saturday 21 March 2026  00:49:05 +0000 (0:00:00.915)       0:01:46.270 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for designate when using single external frontend] ***",
                                "Saturday 21 March 2026  00:49:10 +0000 (0:00:05.051)       0:01:51.321 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for designate] *********************",
                                "Saturday 21 March 2026  00:49:11 +0000 (0:00:01.464)       0:01:52.786 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over designate ProxySQL users config] **********",
                                "Saturday 21 March 2026  00:49:13 +0000 (0:00:01.948)       0:01:54.734 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over designate ProxySQL rules config] **********",
                                "Saturday 21 March 2026  00:49:14 +0000 (0:00:01.243)       0:01:55.977 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : etcd] *****************************************************",
                                "Saturday 21 March 2026  00:49:16 +0000 (0:00:02.132)       0:01:58.109 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : glance] ***************************************************",
                                "Saturday 21 March 2026  00:49:17 +0000 (0:00:00.285)       0:01:58.395 ******** ",
                                "\u001b[0;36mincluded: glance for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over glance haproxy config] *********************",
                                "Saturday 21 March 2026  00:49:18 +0000 (0:00:00.938)       0:01:59.333 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for glance when using single external frontend] ***",
                                "Saturday 21 March 2026  00:49:21 +0000 (0:00:03.828)       0:02:03.162 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for glance] ************************",
                                "Saturday 21 March 2026  00:49:24 +0000 (0:00:02.525)       0:02:05.687 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over glance ProxySQL users config] *************",
                                "Saturday 21 March 2026  00:49:27 +0000 (0:00:02.900)       0:02:08.587 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over glance ProxySQL rules config] *************",
                                "Saturday 21 March 2026  00:49:28 +0000 (0:00:01.150)       0:02:09.738 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : gnocchi] **************************************************",
                                "Saturday 21 March 2026  00:49:30 +0000 (0:00:01.665)       0:02:11.404 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : grafana] **************************************************",
                                "Saturday 21 March 2026  00:49:30 +0000 (0:00:00.258)       0:02:11.662 ******** ",
                                "\u001b[0;36mincluded: grafana for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over grafana haproxy config] ********************",
                                "Saturday 21 March 2026  00:49:31 +0000 (0:00:00.867)       0:02:12.530 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for grafana when using single external frontend] ***",
                                "Saturday 21 March 2026  00:49:34 +0000 (0:00:02.816)       0:02:15.347 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for grafana] ***********************",
                                "Saturday 21 March 2026  00:49:34 +0000 (0:00:00.398)       0:02:15.745 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over grafana ProxySQL users config] ************",
                                "Saturday 21 March 2026  00:49:35 +0000 (0:00:00.791)       0:02:16.536 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over grafana ProxySQL rules config] ************",
                                "Saturday 21 March 2026  00:49:36 +0000 (0:00:01.371)       0:02:17.908 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : heat] *****************************************************",
                                "Saturday 21 March 2026  00:49:38 +0000 (0:00:02.001)       0:02:19.910 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : horizon] **************************************************",
                                "Saturday 21 March 2026  00:49:38 +0000 (0:00:00.250)       0:02:20.160 ******** ",
                                "\u001b[0;36mincluded: horizon for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over horizon haproxy config] ********************",
                                "Saturday 21 March 2026  00:49:39 +0000 (0:00:00.944)       0:02:21.104 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for horizon when using single external frontend] ***",
                                "Saturday 21 March 2026  00:49:43 +0000 (0:00:03.190)       0:02:24.295 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for horizon] ***********************",
                                "Saturday 21 March 2026  00:49:43 +0000 (0:00:00.589)       0:02:24.884 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over horizon ProxySQL users config] ************",
                                "Saturday 21 March 2026  00:49:44 +0000 (0:00:00.860)       0:02:25.745 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over horizon ProxySQL rules config] ************",
                                "Saturday 21 March 2026  00:49:46 +0000 (0:00:01.695)       0:02:27.440 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : influxdb] *************************************************",
                                "Saturday 21 March 2026  00:49:48 +0000 (0:00:01.931)       0:02:29.372 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : ironic] ***************************************************",
                                "Saturday 21 March 2026  00:49:48 +0000 (0:00:00.386)       0:02:29.758 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : keystone] *************************************************",
                                "Saturday 21 March 2026  00:49:48 +0000 (0:00:00.272)       0:02:30.030 ******** ",
                                "\u001b[0;36mincluded: keystone for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over keystone haproxy config] *******************",
                                "Saturday 21 March 2026  00:49:49 +0000 (0:00:01.006)       0:02:31.036 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for keystone when using single external frontend] ***",
                                "Saturday 21 March 2026  00:49:53 +0000 (0:00:03.604)       0:02:34.641 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for keystone] **********************",
                                "Saturday 21 March 2026  00:49:54 +0000 (0:00:00.629)       0:02:35.270 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over keystone ProxySQL users config] ***********",
                                "Saturday 21 March 2026  00:49:55 +0000 (0:00:01.117)       0:02:36.387 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over keystone ProxySQL rules config] ***********",
                                "Saturday 21 March 2026  00:49:56 +0000 (0:00:01.215)       0:02:37.603 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : letsencrypt] **********************************************",
                                "Saturday 21 March 2026  00:49:58 +0000 (0:00:02.056)       0:02:39.660 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : magnum] ***************************************************",
                                "Saturday 21 March 2026  00:49:58 +0000 (0:00:00.280)       0:02:39.941 ******** ",
                                "\u001b[0;36mincluded: magnum for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over magnum haproxy config] *********************",
                                "Saturday 21 March 2026  00:49:59 +0000 (0:00:01.068)       0:02:41.009 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for magnum when using single external frontend] ***",
                                "Saturday 21 March 2026  00:50:02 +0000 (0:00:03.124)       0:02:44.134 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for magnum] ************************",
                                "Saturday 21 March 2026  00:50:03 +0000 (0:00:00.621)       0:02:44.756 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over magnum ProxySQL users config] *************",
                                "Saturday 21 March 2026  00:50:04 +0000 (0:00:01.080)       0:02:45.836 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over magnum ProxySQL rules config] *************",
                                "Saturday 21 March 2026  00:50:06 +0000 (0:00:01.358)       0:02:47.195 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : manila] ***************************************************",
                                "Saturday 21 March 2026  00:50:08 +0000 (0:00:02.095)       0:02:49.290 ******** ",
                                "\u001b[0;36mincluded: manila for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over manila haproxy config] *********************",
                                "Saturday 21 March 2026  00:50:09 +0000 (0:00:01.054)       0:02:50.345 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for manila when using single external frontend] ***",
                                "Saturday 21 March 2026  00:50:13 +0000 (0:00:04.135)       0:02:54.481 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for manila] ************************",
                                "Saturday 21 March 2026  00:50:14 +0000 (0:00:00.844)       0:02:55.325 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over manila ProxySQL users config] *************",
                                "Saturday 21 March 2026  00:50:15 +0000 (0:00:01.076)       0:02:56.402 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over manila ProxySQL rules config] *************",
                                "Saturday 21 March 2026  00:50:16 +0000 (0:00:01.431)       0:02:57.833 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : mariadb] **************************************************",
                                "Saturday 21 March 2026  00:50:18 +0000 (0:00:02.182)       0:03:00.016 ******** ",
                                "\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [mariadb : Ensure mysql monitor user exist] *******************************",
                                "Saturday 21 March 2026  00:50:20 +0000 (0:00:01.281)       0:03:01.298 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over mariadb haproxy config] ********************",
                                "Saturday 21 March 2026  00:50:23 +0000 (0:00:03.128)       0:03:04.427 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for mariadb when using single external frontend] ***",
                                "Saturday 21 March 2026  00:50:25 +0000 (0:00:02.199)       0:03:06.627 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for mariadb] ***********************",
                                "Saturday 21 March 2026  00:50:27 +0000 (0:00:02.163)       0:03:08.790 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over mariadb ProxySQL users config] ************",
                                "Saturday 21 March 2026  00:50:29 +0000 (0:00:02.027)       0:03:10.817 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over mariadb ProxySQL rules config] ************",
                                "Saturday 21 March 2026  00:50:31 +0000 (0:00:01.840)       0:03:12.658 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : masakari] *************************************************",
                                "Saturday 21 March 2026  00:50:32 +0000 (0:00:01.397)       0:03:14.056 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : memcached] ************************************************",
                                "Saturday 21 March 2026  00:50:33 +0000 (0:00:00.272)       0:03:14.328 ******** ",
                                "\u001b[0;36mincluded: memcached for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over memcached haproxy config] ******************",
                                "Saturday 21 March 2026  00:50:34 +0000 (0:00:01.179)       0:03:15.508 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for memcached when using single external frontend] ***",
                                "Saturday 21 March 2026  00:50:35 +0000 (0:00:01.427)       0:03:16.936 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for memcached] *********************",
                                "Saturday 21 March 2026  00:50:36 +0000 (0:00:00.349)       0:03:17.285 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over memcached ProxySQL users config] **********",
                                "Saturday 21 March 2026  00:50:36 +0000 (0:00:00.734)       0:03:18.020 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over memcached ProxySQL rules config] **********",
                                "Saturday 21 March 2026  00:50:37 +0000 (0:00:00.432)       0:03:18.453 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : mistral] **************************************************",
                                "Saturday 21 March 2026  00:50:38 +0000 (0:00:01.074)       0:03:19.527 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : neutron] **************************************************",
                                "Saturday 21 March 2026  00:50:38 +0000 (0:00:00.264)       0:03:19.792 ******** ",
                                "\u001b[0;36mincluded: neutron for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over neutron haproxy config] ********************",
                                "Saturday 21 March 2026  00:50:39 +0000 (0:00:01.273)       0:03:21.066 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for neutron when using single external frontend] ***",
                                "Saturday 21 March 2026  00:50:43 +0000 (0:00:03.957)       0:03:25.023 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for neutron] ***********************",
                                "Saturday 21 March 2026  00:50:45 +0000 (0:00:01.820)       0:03:26.844 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over neutron ProxySQL users config] ************",
                                "Saturday 21 March 2026  00:50:47 +0000 (0:00:01.427)       0:03:28.271 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over neutron ProxySQL rules config] ************",
                                "Saturday 21 March 2026  00:50:48 +0000 (0:00:01.339)       0:03:29.610 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : placement] ************************************************",
                                "Saturday 21 March 2026  00:50:50 +0000 (0:00:01.921)       0:03:31.532 ******** ",
                                "\u001b[0;36mincluded: placement for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over placement haproxy config] ******************",
                                "Saturday 21 March 2026  00:50:51 +0000 (0:00:01.231)       0:03:32.763 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for placement when using single external frontend] ***",
                                "Saturday 21 March 2026  00:50:54 +0000 (0:00:02.896)       0:03:35.660 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for placement] *********************",
                                "Saturday 21 March 2026  00:50:54 +0000 (0:00:00.428)       0:03:36.088 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over placement ProxySQL users config] **********",
                                "Saturday 21 March 2026  00:50:55 +0000 (0:00:01.058)       0:03:37.146 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over placement ProxySQL rules config] **********",
                                "Saturday 21 March 2026  00:50:57 +0000 (0:00:01.359)       0:03:38.506 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : nova] *****************************************************",
                                "Saturday 21 March 2026  00:50:59 +0000 (0:00:01.931)       0:03:40.437 ******** ",
                                "\u001b[0;36mincluded: nova for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over nova haproxy config] ***********************",
                                "Saturday 21 March 2026  00:51:00 +0000 (0:00:01.314)       0:03:41.752 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for nova when using single external frontend] ***",
                                "Saturday 21 March 2026  00:51:04 +0000 (0:00:03.693)       0:03:45.446 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for nova] **************************",
                                "Saturday 21 March 2026  00:51:04 +0000 (0:00:00.564)       0:03:46.010 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova ProxySQL users config] ***************",
                                "Saturday 21 March 2026  00:51:05 +0000 (0:00:00.818)       0:03:46.829 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova ProxySQL rules config] ***************",
                                "Saturday 21 March 2026  00:51:07 +0000 (0:00:01.647)       0:03:48.476 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : nova-cell] ************************************************",
                                "Saturday 21 March 2026  00:51:09 +0000 (0:00:02.002)       0:03:50.479 ******** ",
                                "\u001b[0;36mincluded: nova-cell for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova-cell : Configure loadbalancer for nova-novncproxy] ******************",
                                "Saturday 21 March 2026  00:51:10 +0000 (0:00:01.248)       0:03:51.727 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-novncproxy)\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over nova-cell:nova-novncproxy haproxy config] ***",
                                "Saturday 21 March 2026  00:51:11 +0000 (0:00:01.122)       0:03:52.850 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for nova-cell:nova-novncproxy when using single external frontend] ***",
                                "Saturday 21 March 2026  00:51:15 +0000 (0:00:03.420)       0:03:56.271 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for nova-cell:nova-novncproxy] *****",
                                "Saturday 21 March 2026  00:51:16 +0000 (0:00:01.502)       0:03:57.773 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********",
                                "Saturday 21 March 2026  00:51:18 +0000 (0:00:01.880)       0:03:59.654 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********",
                                "Saturday 21 March 2026  00:51:20 +0000 (0:00:02.244)       0:04:01.898 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Configure loadbalancer for nova-spicehtml5proxy] *************",
                                "Saturday 21 March 2026  00:51:23 +0000 (0:00:02.605)       0:04:04.503 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-spicehtml5proxy)\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over nova-cell:nova-spicehtml5proxy haproxy config] ***",
                                "Saturday 21 March 2026  00:51:24 +0000 (0:00:00.759)       0:04:05.262 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for nova-cell:nova-spicehtml5proxy when using single external frontend] ***",
                                "Saturday 21 March 2026  00:51:25 +0000 (0:00:01.112)       0:04:06.375 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for nova-cell:nova-spicehtml5proxy] ***",
                                "Saturday 21 March 2026  00:51:26 +0000 (0:00:01.327)       0:04:07.702 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********",
                                "Saturday 21 March 2026  00:51:27 +0000 (0:00:01.299)       0:04:09.001 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********",
                                "Saturday 21 March 2026  00:51:30 +0000 (0:00:02.259)       0:04:11.261 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [nova-cell : Configure loadbalancer for nova-serialproxy] *****************",
                                "Saturday 21 March 2026  00:51:32 +0000 (0:00:02.861)       0:04:14.122 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-serialproxy)\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over nova-cell:nova-serialproxy haproxy config] ***",
                                "Saturday 21 March 2026  00:51:33 +0000 (0:00:00.849)       0:04:14.971 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for nova-cell:nova-serialproxy when using single external frontend] ***",
                                "Saturday 21 March 2026  00:51:34 +0000 (0:00:01.146)       0:04:16.118 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for nova-cell:nova-serialproxy] ****",
                                "Saturday 21 March 2026  00:51:36 +0000 (0:00:01.116)       0:04:17.235 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********",
                                "Saturday 21 March 2026  00:51:37 +0000 (0:00:01.287)       0:04:18.522 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********",
                                "Saturday 21 March 2026  00:51:39 +0000 (0:00:02.458)       0:04:20.981 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : octavia] **************************************************",
                                "Saturday 21 March 2026  00:51:42 +0000 (0:00:02.949)       0:04:23.930 ******** ",
                                "\u001b[0;36mincluded: octavia for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over octavia haproxy config] ********************",
                                "Saturday 21 March 2026  00:51:43 +0000 (0:00:01.220)       0:04:25.151 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for octavia when using single external frontend] ***",
                                "Saturday 21 March 2026  00:51:47 +0000 (0:00:03.471)       0:04:28.622 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for octavia] ***********************",
                                "Saturday 21 March 2026  00:51:48 +0000 (0:00:00.877)       0:04:29.500 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over octavia ProxySQL users config] ************",
                                "Saturday 21 March 2026  00:51:49 +0000 (0:00:00.920)       0:04:30.420 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over octavia ProxySQL rules config] ************",
                                "Saturday 21 March 2026  00:51:50 +0000 (0:00:01.318)       0:04:31.739 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : opensearch] ***********************************************",
                                "Saturday 21 March 2026  00:51:52 +0000 (0:00:02.090)       0:04:33.830 ******** ",
                                "\u001b[0;36mincluded: opensearch for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over opensearch haproxy config] *****************",
                                "Saturday 21 March 2026  00:51:54 +0000 (0:00:01.597)       0:04:35.427 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for opensearch when using single external frontend] ***",
                                "Saturday 21 March 2026  00:51:58 +0000 (0:00:04.726)       0:04:40.153 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for opensearch] ********************",
                                "Saturday 21 March 2026  00:51:59 +0000 (0:00:00.809)       0:04:40.962 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over opensearch ProxySQL users config] *********",
                                "Saturday 21 March 2026  00:52:00 +0000 (0:00:01.056)       0:04:42.019 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over opensearch ProxySQL rules config] *********",
                                "Saturday 21 March 2026  00:52:01 +0000 (0:00:00.417)       0:04:42.437 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : prometheus] ***********************************************",
                                "Saturday 21 March 2026  00:52:02 +0000 (0:00:01.126)       0:04:43.563 ******** ",
                                "\u001b[0;36mincluded: prometheus for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over prometheus haproxy config] *****************",
                                "Saturday 21 March 2026  00:52:03 +0000 (0:00:01.541)       0:04:45.105 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for prometheus when using single external frontend] ***",
                                "Saturday 21 March 2026  00:52:08 +0000 (0:00:04.081)       0:04:49.187 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for prometheus] ********************",
                                "Saturday 21 March 2026  00:52:08 +0000 (0:00:00.761)       0:04:49.948 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over prometheus ProxySQL users config] *********",
                                "Saturday 21 March 2026  00:52:09 +0000 (0:00:01.047)       0:04:50.996 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over prometheus ProxySQL rules config] *********",
                                "Saturday 21 March 2026  00:52:10 +0000 (0:00:00.428)       0:04:51.424 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : rabbitmq] *************************************************",
                                "Saturday 21 March 2026  00:52:11 +0000 (0:00:01.111)       0:04:52.536 ******** ",
                                "\u001b[0;36mincluded: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over rabbitmq haproxy config] *******************",
                                "Saturday 21 March 2026  00:52:12 +0000 (0:00:01.322)       0:04:53.858 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for rabbitmq when using single external frontend] ***",
                                "Saturday 21 March 2026  00:52:14 +0000 (0:00:02.105)       0:04:55.964 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for rabbitmq] **********************",
                                "Saturday 21 March 2026  00:52:15 +0000 (0:00:00.395)       0:04:56.359 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over rabbitmq ProxySQL users config] ***********",
                                "Saturday 21 March 2026  00:52:15 +0000 (0:00:00.577)       0:04:56.937 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over rabbitmq ProxySQL rules config] ***********",
                                "Saturday 21 March 2026  00:52:16 +0000 (0:00:00.656)       0:04:57.593 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : skyline] **************************************************",
                                "Saturday 21 March 2026  00:52:17 +0000 (0:00:01.117)       0:04:58.711 ******** ",
                                "\u001b[0;36mincluded: skyline for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over skyline haproxy config] ********************",
                                "Saturday 21 March 2026  00:52:18 +0000 (0:00:01.405)       0:05:00.116 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for skyline when using single external frontend] ***",
                                "Saturday 21 March 2026  00:52:24 +0000 (0:00:05.240)       0:05:05.357 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for skyline] ***********************",
                                "Saturday 21 March 2026  00:52:25 +0000 (0:00:00.830)       0:05:06.188 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over skyline ProxySQL users config] ************",
                                "Saturday 21 March 2026  00:52:25 +0000 (0:00:00.833)       0:05:07.021 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over skyline ProxySQL rules config] ************",
                                "Saturday 21 March 2026  00:52:27 +0000 (0:00:01.330)       0:05:08.352 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : swift] ****************************************************",
                                "Saturday 21 March 2026  00:52:29 +0000 (0:00:02.079)       0:05:10.432 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : tacker] ***************************************************",
                                "Saturday 21 March 2026  00:52:29 +0000 (0:00:00.477)       0:05:10.910 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : trove] ****************************************************",
                                "Saturday 21 March 2026  00:52:29 +0000 (0:00:00.268)       0:05:11.178 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : venus] ****************************************************",
                                "Saturday 21 March 2026  00:52:30 +0000 (0:00:00.282)       0:05:11.461 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : watcher] **************************************************",
                                "Saturday 21 March 2026  00:52:30 +0000 (0:00:00.264)       0:05:11.726 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : zun] ******************************************************",
                                "Saturday 21 March 2026  00:52:31 +0000 (0:00:00.467)       0:05:12.193 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Check IP addresses on the API interface] *******",
                                "Saturday 21 March 2026  00:52:31 +0000 (0:00:00.491)       0:05:12.685 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Group HA nodes by status] **********************",
                                "Saturday 21 March 2026  00:52:32 +0000 (0:00:00.663)       0:05:13.348 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop backup keepalived container] **************",
                                "Saturday 21 March 2026  00:52:32 +0000 (0:00:00.496)       0:05:13.845 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop backup haproxy container] *****************",
                                "Saturday 21 March 2026  00:52:33 +0000 (0:00:00.901)       0:05:14.746 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop backup proxysql container] ****************",
                                "Saturday 21 March 2026  00:52:34 +0000 (0:00:01.030)       0:05:15.777 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start backup haproxy container] ****************",
                                "Saturday 21 March 2026  00:52:35 +0000 (0:00:00.969)       0:05:16.746 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for backup haproxy to start] **************",
                                "Saturday 21 March 2026  00:52:39 +0000 (0:00:04.268)       0:05:21.015 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start backup proxysql container] ***************",
                                "Saturday 21 March 2026  00:52:42 +0000 (0:00:03.032)       0:05:24.047 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for backup proxysql to start] *************",
                                "Saturday 21 March 2026  00:52:51 +0000 (0:00:08.295)       0:05:32.343 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start backup keepalived container] *************",
                                "Saturday 21 March 2026  00:52:54 +0000 (0:00:03.751)       0:05:36.094 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop master haproxy container] *****************",
                                "Saturday 21 March 2026  00:52:59 +0000 (0:00:04.599)       0:05:40.694 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop master proxysql container] ****************",
                                "Saturday 21 March 2026  00:53:00 +0000 (0:00:00.534)       0:05:41.228 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop master keepalived container] **************",
                                "Saturday 21 March 2026  00:53:00 +0000 (0:00:00.312)       0:05:41.541 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start master haproxy container] ****************",
                                "Saturday 21 March 2026  00:53:00 +0000 (0:00:00.335)       0:05:41.876 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start master proxysql container] ***************",
                                "Saturday 21 March 2026  00:53:01 +0000 (0:00:00.315)       0:05:42.192 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start master keepalived container] *************",
                                "Saturday 21 March 2026  00:53:01 +0000 (0:00:00.546)       0:05:42.739 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for haproxy to listen on VIP] *************",
                                "Saturday 21 March 2026  00:53:01 +0000 (0:00:00.302)       0:05:43.041 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for proxysql to listen on VIP] ************",
                                "Saturday 21 March 2026  00:53:06 +0000 (0:00:04.728)       0:05:47.770 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=123 \u001b[0m \u001b[0;33mchanged=76  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=97  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=122 \u001b[0m \u001b[0;33mchanged=76  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=97  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=122 \u001b[0m \u001b[0;33mchanged=76  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=97  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:53:07 +0000 (0:00:00.868)       0:05:48.639 ******** ",
                                "=============================================================================== ",
                                "loadbalancer : Start backup proxysql container -------------------------- 8.30s",
                                "loadbalancer : Copying over proxysql config ----------------------------- 5.49s",
                                "haproxy-config : Copying over skyline haproxy config -------------------- 5.24s",
                                "haproxy-config : Copying over designate haproxy config ------------------ 5.05s",
                                "loadbalancer : Wait for haproxy to listen on VIP ------------------------ 4.73s",
                                "haproxy-config : Copying over opensearch haproxy config ----------------- 4.73s",
                                "loadbalancer : Start backup keepalived container ------------------------ 4.60s",
                                "haproxy-config : Copying over barbican haproxy config ------------------- 4.28s",
                                "loadbalancer : Start backup haproxy container --------------------------- 4.27s",
                                "haproxy-config : Copying over manila haproxy config --------------------- 4.14s",
                                "haproxy-config : Copying over prometheus haproxy config ----------------- 4.08s",
                                "haproxy-config : Copying over neutron haproxy config -------------------- 3.96s",
                                "haproxy-config : Copying over glance haproxy config --------------------- 3.83s",
                                "haproxy-config : Copying over cinder haproxy config --------------------- 3.75s",
                                "loadbalancer : Wait for backup proxysql to start ------------------------ 3.75s",
                                "haproxy-config : Copying over nova haproxy config ----------------------- 3.69s",
                                "haproxy-config : Copying over keystone haproxy config ------------------- 3.60s",
                                "haproxy-config : Copying over octavia haproxy config -------------------- 3.47s",
                                "haproxy-config : Copying over nova-cell:nova-novncproxy haproxy config --- 3.42s",
                                "haproxy-config : Copying over aodh haproxy config ----------------------- 3.42s",
                                "\u001b[32m2026-03-21 00:53:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:53:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:54:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b3f66793-968d-4cfe-b8b7-2ae147b8063e is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 51831d5c-a9eb-44f2-b3ad-4ceca7d481d6 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  00:53:11 +0000 (0:00:00.284)       0:00:00.284 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  00:53:11 +0000 (0:00:00.269)       0:00:00.553 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_opensearch_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_opensearch_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_opensearch_True)\u001b[0m",
                                "",
                                "PLAY [Apply role opensearch] ***************************************************",
                                "",
                                "TASK [opensearch : include_tasks] **********************************************",
                                "Saturday 21 March 2026  00:53:11 +0000 (0:00:00.279)       0:00:00.833 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [opensearch : Setting sysctl values] **************************************",
                                "Saturday 21 March 2026  00:53:12 +0000 (0:00:00.583)       0:00:01.416 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "",
                                "TASK [opensearch : Ensuring config directories exist] **************************",
                                "Saturday 21 March 2026  00:53:13 +0000 (0:00:00.937)       0:00:02.354 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "",
                                "TASK [opensearch : include_tasks] **********************************************",
                                "Saturday 21 March 2026  00:53:14 +0000 (0:00:01.284)       0:00:03.638 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : opensearch | Copying over extra CA certificates] *****",
                                "Saturday 21 March 2026  00:53:15 +0000 (0:00:00.585)       0:00:04.224 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : opensearch | Copying over backend internal TLS certificate] ***",
                                "Saturday 21 March 2026  00:53:18 +0000 (0:00:03.593)       0:00:07.817 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : opensearch | Copying over backend internal TLS key] ***",
                                "Saturday 21 March 2026  00:53:19 +0000 (0:00:00.777)       0:00:08.595 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [opensearch : Copying over config.json files for services] ****************",
                                "Saturday 21 March 2026  00:53:20 +0000 (0:00:00.769)       0:00:09.364 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "",
                                "TASK [opensearch : Copying over opensearch service config file] ****************",
                                "Saturday 21 March 2026  00:53:22 +0000 (0:00:02.721)       0:00:12.085 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [opensearch : Copying over opensearch-dashboards config file] *************",
                                "Saturday 21 March 2026  00:53:25 +0000 (0:00:02.984)       0:00:15.069 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [opensearch : Check opensearch containers] ********************************",
                                "Saturday 21 March 2026  00:53:27 +0000 (0:00:01.500)       0:00:16.570 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "",
                                "TASK [opensearch : include_tasks] **********************************************",
                                "Saturday 21 March 2026  00:53:29 +0000 (0:00:02.237)       0:00:18.808 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [opensearch : Flush handlers] *********************************************",
                                "Saturday 21 March 2026  00:53:30 +0000 (0:00:00.427)       0:00:19.235 ******** ",
                                "",
                                "TASK [opensearch : Flush handlers] *********************************************",
                                "Saturday 21 March 2026  00:53:30 +0000 (0:00:00.066)       0:00:19.302 ******** ",
                                "",
                                "TASK [opensearch : Flush handlers] *********************************************",
                                "Saturday 21 March 2026  00:53:30 +0000 (0:00:00.065)       0:00:19.367 ******** ",
                                "",
                                "RUNNING HANDLER [opensearch : Disable shard allocation] ************************",
                                "Saturday 21 March 2026  00:53:30 +0000 (0:00:00.067)       0:00:19.435 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [opensearch : Perform a flush] *********************************",
                                "Saturday 21 March 2026  00:53:30 +0000 (0:00:00.243)       0:00:19.678 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [opensearch : Restart opensearch container] ********************",
                                "Saturday 21 March 2026  00:53:30 +0000 (0:00:00.237)       0:00:19.916 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [opensearch : Restart opensearch-dashboards container] *********",
                                "Saturday 21 March 2026  00:54:27 +0000 (0:00:56.884)       0:01:16.801 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [opensearch : include_tasks] **********************************************",
                                "Saturday 21 March 2026  00:55:34 +0000 (0:01:07.297)       0:02:24.098 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/post-config.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [opensearch : Wait for OpenSearch to become ready] ************************",
                                "Saturday 21 March 2026  00:55:35 +0000 (0:00:00.666)       0:02:24.765 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [opensearch : Wait for OpenSearch cluster to become healthy] **************",
                                "Saturday 21 March 2026  00:55:37 +0000 (0:00:02.292)       0:02:27.057 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [opensearch : Check if a log retention policy exists] *********************",
                                "Saturday 21 March 2026  00:55:39 +0000 (0:00:01.932)       0:02:28.989 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [opensearch : Create new log retention policy] ****************************",
                                "Saturday 21 March 2026  00:55:41 +0000 (0:00:01.973)       0:02:30.963 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [opensearch : Apply retention policy to existing indices] *****************",
                                "Saturday 21 March 2026  00:55:44 +0000 (0:00:02.273)       0:02:33.237 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:55:46 +0000 (0:00:02.166)       0:02:35.403 ******** ",
                                "=============================================================================== ",
                                "opensearch : Restart opensearch-dashboards container ------------------- 67.30s",
                                "opensearch : Restart opensearch container ------------------------------ 56.88s",
                                "service-cert-copy : opensearch | Copying over extra CA certificates ----- 3.59s",
                                "opensearch : Copying over opensearch service config file ---------------- 2.98s",
                                "opensearch : Copying over config.json files for services ---------------- 2.72s",
                                "opensearch : Wait for OpenSearch to become ready ------------------------ 2.29s",
                                "opensearch : Create new log retention policy ---------------------------- 2.27s",
                                "opensearch : Check opensearch containers -------------------------------- 2.24s",
                                "opensearch : Apply retention policy to existing indices ----------------- 2.17s",
                                "opensearch : Check if a log retention policy exists --------------------- 1.97s",
                                "opensearch : Wait for OpenSearch cluster to become healthy -------------- 1.93s",
                                "opensearch : Copying over opensearch-dashboards config file ------------- 1.50s",
                                "opensearch : Ensuring config directories exist -------------------------- 1.28s",
                                "opensearch : Setting sysctl values -------------------------------------- 0.94s",
                                "service-cert-copy : opensearch | Copying over backend internal TLS certificate --- 0.78s",
                                "service-cert-copy : opensearch | Copying over backend internal TLS key --- 0.77s",
                                "opensearch : include_tasks ---------------------------------------------- 0.67s",
                                "opensearch : include_tasks ---------------------------------------------- 0.59s",
                                "opensearch : include_tasks ---------------------------------------------- 0.58s",
                                "opensearch : include_tasks ---------------------------------------------- 0.43s",
                                "",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Prepare deployment of Ceph services] *************************************",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Saturday 21 March 2026  00:44:39 +0000 (0:00:00.774)       0:00:00.774 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Saturday 21 March 2026  00:44:40 +0000 (0:00:01.331)       0:00:02.106 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Saturday 21 March 2026  00:44:42 +0000 (0:00:02.059)       0:00:04.165 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Saturday 21 March 2026  00:44:43 +0000 (0:00:00.840)       0:00:05.006 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Saturday 21 March 2026  00:44:45 +0000 (0:00:01.496)       0:00:06.502 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Saturday 21 March 2026  00:44:46 +0000 (0:00:01.230)       0:00:07.733 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Saturday 21 March 2026  00:44:47 +0000 (0:00:00.700)       0:00:08.433 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Saturday 21 March 2026  00:44:47 +0000 (0:00:00.921)       0:00:09.355 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Saturday 21 March 2026  00:44:48 +0000 (0:00:00.926)       0:00:10.282 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Saturday 21 March 2026  00:44:50 +0000 (0:00:01.173)       0:00:11.455 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Saturday 21 March 2026  00:44:50 +0000 (0:00:00.743)       0:00:12.198 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Saturday 21 March 2026  00:44:51 +0000 (0:00:01.166)       0:00:13.365 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Saturday 21 March 2026  00:44:54 +0000 (0:00:02.872)       0:00:16.237 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Saturday 21 March 2026  00:44:55 +0000 (0:00:00.627)       0:00:16.864 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Saturday 21 March 2026  00:44:57 +0000 (0:00:01.603)       0:00:18.467 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Saturday 21 March 2026  00:44:57 +0000 (0:00:00.246)       0:00:18.714 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-21 00:44:52.964412', 'end': '2026-03-21 00:44:53.064620', 'delta': '0:00:00.100208', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-21 00:44:53.564597', 'end': '2026-03-21 00:44:53.659889', 'delta': '0:00:00.095292', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-21 00:44:54.311062', 'end': '2026-03-21 00:44:54.406291', 'delta': '0:00:00.095229', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Saturday 21 March 2026  00:44:57 +0000 (0:00:00.693)       0:00:19.407 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Saturday 21 March 2026  00:45:00 +0000 (0:00:02.996)       0:00:22.404 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Saturday 21 March 2026  00:45:01 +0000 (0:00:00.878)       0:00:23.282 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Saturday 21 March 2026  00:45:03 +0000 (0:00:01.636)       0:00:24.919 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Saturday 21 March 2026  00:45:04 +0000 (0:00:01.023)       0:00:25.943 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Saturday 21 March 2026  00:45:05 +0000 (0:00:01.172)       0:00:27.115 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Saturday 21 March 2026  00:45:05 +0000 (0:00:00.285)       0:00:27.400 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Saturday 21 March 2026  00:45:06 +0000 (0:00:00.448)       0:00:27.848 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Saturday 21 March 2026  00:45:07 +0000 (0:00:01.061)       0:00:28.910 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Saturday 21 March 2026  00:45:08 +0000 (0:00:00.754)       0:00:29.664 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Saturday 21 March 2026  00:45:09 +0000 (0:00:00.871)       0:00:30.535 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Saturday 21 March 2026  00:45:09 +0000 (0:00:00.831)       0:00:31.367 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Saturday 21 March 2026  00:45:10 +0000 (0:00:00.713)       0:00:32.080 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Saturday 21 March 2026  00:45:12 +0000 (0:00:01.479)       0:00:33.559 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Saturday 21 March 2026  00:45:12 +0000 (0:00:00.656)       0:00:34.216 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--64dd100a--6bcc--5763--b557--0732c79bfd55-osd--block--64dd100a--6bcc--5763--b557--0732c79bfd55', 'dm-uuid-LVM-MWn87OlXxZ0hjMuGG89puNwiFbMviFcm1xAms0qPz8LYU8Ut0AGzeF9WoI61D5NM'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--5368e13a--d21f--5912--91bd--da537815d9e8-osd--block--5368e13a--d21f--5912--91bd--da537815d9e8', 'dm-uuid-LVM-xgLPNxLNUSwUnkWnyYv6gd71t6QIH1sH9rPUXF3date4d4aBKBICcuqoW7mNU6TI'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part1', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part14', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part15', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part16', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'holders': ['ceph--64dd100a--6bcc--5763--b557--0732c79bfd55-osd--block--64dd100a--6bcc--5763--b557--0732c79bfd55'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-HiPrG3-9qzo-WJZ0-1Smx-Bi1t-ReBz-6cdxta', 'scsi-0QEMU_QEMU_HARDDISK_f08311eb-e2ad-4eb7-8da9-ac6d899a2409', 'scsi-SQEMU_QEMU_HARDDISK_f08311eb-e2ad-4eb7-8da9-ac6d899a2409'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'holders': ['ceph--5368e13a--d21f--5912--91bd--da537815d9e8-osd--block--5368e13a--d21f--5912--91bd--da537815d9e8'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-m0F2nF-j9UB-mZfT-bRv8-xW1c-EVch-ZfDvHt', 'scsi-0QEMU_QEMU_HARDDISK_e03b171b-b829-41a0-a0b6-cc46d85f013d', 'scsi-SQEMU_QEMU_HARDDISK_e03b171b-b829-41a0-a0b6-cc46d85f013d'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_9fa3c153-0e2a-45ef-8e48-e0095eb937ad', 'scsi-SQEMU_QEMU_HARDDISK_9fa3c153-0e2a-45ef-8e48-e0095eb937ad'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-21-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--ed04aa0b--4f5a--523b--a5d2--05072a1b260d-osd--block--ed04aa0b--4f5a--523b--a5d2--05072a1b260d', 'dm-uuid-LVM-viMZkgfP3jZUgTAo0bSRpd40fLtFkJJmO8XW5mmi1aRChVrq8X3DDkwGl8XRb5at'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--45b09425--beda--50b4--a83b--1fc50b5adca4-osd--block--45b09425--beda--50b4--a83b--1fc50b5adca4', 'dm-uuid-LVM-XzfTGfmLHoftdDBti3ot8l7sbRObO8yALj7bfZRysK7QQ9WjzIgwF2E0nO1Pc63W'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part1', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part14', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part15', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part16', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'holders': ['ceph--ed04aa0b--4f5a--523b--a5d2--05072a1b260d-osd--block--ed04aa0b--4f5a--523b--a5d2--05072a1b260d'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-nra7ru-O0w6-dM26-lz5o-2arZ-rXsE-lANagX', 'scsi-0QEMU_QEMU_HARDDISK_4e74a544-8b3a-4a05-98a1-b7cc13f13c18', 'scsi-SQEMU_QEMU_HARDDISK_4e74a544-8b3a-4a05-98a1-b7cc13f13c18'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'holders': ['ceph--45b09425--beda--50b4--a83b--1fc50b5adca4-osd--block--45b09425--beda--50b4--a83b--1fc50b5adca4'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-XQTe4H-QTjK-DLSs-Qjj6-RKjC-da4T-xXChra', 'scsi-0QEMU_QEMU_HARDDISK_666e73fe-87f5-4f65-9ff2-95a038c8b18d', 'scsi-SQEMU_QEMU_HARDDISK_666e73fe-87f5-4f65-9ff2-95a038c8b18d'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ae4aabb-0174-4999-a1d0-4bf78cbb2965', 'scsi-SQEMU_QEMU_HARDDISK_8ae4aabb-0174-4999-a1d0-4bf78cbb2965'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-14-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--ede169e1--d974--55ed--a938--c303cb89cb5b-osd--block--ede169e1--d974--55ed--a938--c303cb89cb5b', 'dm-uuid-LVM-bEBXWAqtQdfT6xKYbL0dF10We9CLdKgKh0pf0vpR2lQnL405kp4ndXs76m3o0J5Y'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a-osd--block--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a', 'dm-uuid-LVM-G1LwYmzPCvwLZRWX1y5oYSHlzZdIxlOhfbzzvktDUwfRIgHaiaw6r0MV8HVh1TLk'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part1', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part14', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part15', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part16', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'holders': ['ceph--ede169e1--d974--55ed--a938--c303cb89cb5b-osd--block--ede169e1--d974--55ed--a938--c303cb89cb5b'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-HVhEyD-n8kM-PvHm-CUDD-tuRZ-igjC-Nzour2', 'scsi-0QEMU_QEMU_HARDDISK_f76ae8f0-65cc-4a62-b19c-23f8a40fbde6', 'scsi-SQEMU_QEMU_HARDDISK_f76ae8f0-65cc-4a62-b19c-23f8a40fbde6'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'holders': ['ceph--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a-osd--block--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-YifJdB-JYWE-HhRV-mF1g-0k83-hzWG-ctJZWy', 'scsi-0QEMU_QEMU_HARDDISK_d7bb81ba-6491-4b5a-bb82-f0b49314eac0', 'scsi-SQEMU_QEMU_HARDDISK_d7bb81ba-6491-4b5a-bb82-f0b49314eac0'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_cdffe5f1-7f0d-460e-a1db-d6846d226144', 'scsi-SQEMU_QEMU_HARDDISK_cdffe5f1-7f0d-460e-a1db-d6846d226144'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-28-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f', 'scsi-SQEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part1', 'scsi-SQEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part14', 'scsi-SQEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part15', 'scsi-SQEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part16', 'scsi-SQEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-10-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5', 'scsi-SQEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part1', 'scsi-SQEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part14', 'scsi-SQEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part15', 'scsi-SQEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part16', 'scsi-SQEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-23-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32', 'scsi-SQEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part1', 'scsi-SQEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part14', 'scsi-SQEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part15', 'scsi-SQEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part16', 'scsi-SQEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-19-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Saturday 21 March 2026  00:45:14 +0000 (0:00:01.751)       0:00:35.968 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--64dd100a--6bcc--5763--b557--0732c79bfd55-osd--block--64dd100a--6bcc--5763--b557--0732c79bfd55', 'dm-uuid-LVM-MWn87OlXxZ0hjMuGG89puNwiFbMviFcm1xAms0qPz8LYU8Ut0AGzeF9WoI61D5NM'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--5368e13a--d21f--5912--91bd--da537815d9e8-osd--block--5368e13a--d21f--5912--91bd--da537815d9e8', 'dm-uuid-LVM-xgLPNxLNUSwUnkWnyYv6gd71t6QIH1sH9rPUXF3date4d4aBKBICcuqoW7mNU6TI'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--ed04aa0b--4f5a--523b--a5d2--05072a1b260d-osd--block--ed04aa0b--4f5a--523b--a5d2--05072a1b260d', 'dm-uuid-LVM-viMZkgfP3jZUgTAo0bSRpd40fLtFkJJmO8XW5mmi1aRChVrq8X3DDkwGl8XRb5at'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--45b09425--beda--50b4--a83b--1fc50b5adca4-osd--block--45b09425--beda--50b4--a83b--1fc50b5adca4', 'dm-uuid-LVM-XzfTGfmLHoftdDBti3ot8l7sbRObO8yALj7bfZRysK7QQ9WjzIgwF2E0nO1Pc63W'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part1', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part14', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part15', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part16', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--ed04aa0b--4f5a--523b--a5d2--05072a1b260d-osd--block--ed04aa0b--4f5a--523b--a5d2--05072a1b260d'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-nra7ru-O0w6-dM26-lz5o-2arZ-rXsE-lANagX', 'scsi-0QEMU_QEMU_HARDDISK_4e74a544-8b3a-4a05-98a1-b7cc13f13c18', 'scsi-SQEMU_QEMU_HARDDISK_4e74a544-8b3a-4a05-98a1-b7cc13f13c18'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--ede169e1--d974--55ed--a938--c303cb89cb5b-osd--block--ede169e1--d974--55ed--a938--c303cb89cb5b', 'dm-uuid-LVM-bEBXWAqtQdfT6xKYbL0dF10We9CLdKgKh0pf0vpR2lQnL405kp4ndXs76m3o0J5Y'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--45b09425--beda--50b4--a83b--1fc50b5adca4-osd--block--45b09425--beda--50b4--a83b--1fc50b5adca4'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-XQTe4H-QTjK-DLSs-Qjj6-RKjC-da4T-xXChra', 'scsi-0QEMU_QEMU_HARDDISK_666e73fe-87f5-4f65-9ff2-95a038c8b18d', 'scsi-SQEMU_QEMU_HARDDISK_666e73fe-87f5-4f65-9ff2-95a038c8b18d'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ae4aabb-0174-4999-a1d0-4bf78cbb2965', 'scsi-SQEMU_QEMU_HARDDISK_8ae4aabb-0174-4999-a1d0-4bf78cbb2965'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part1', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part14', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part15', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part16', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a-osd--block--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a', 'dm-uuid-LVM-G1LwYmzPCvwLZRWX1y5oYSHlzZdIxlOhfbzzvktDUwfRIgHaiaw6r0MV8HVh1TLk'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-14-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--64dd100a--6bcc--5763--b557--0732c79bfd55-osd--block--64dd100a--6bcc--5763--b557--0732c79bfd55'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-HiPrG3-9qzo-WJZ0-1Smx-Bi1t-ReBz-6cdxta', 'scsi-0QEMU_QEMU_HARDDISK_f08311eb-e2ad-4eb7-8da9-ac6d899a2409', 'scsi-SQEMU_QEMU_HARDDISK_f08311eb-e2ad-4eb7-8da9-ac6d899a2409'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--5368e13a--d21f--5912--91bd--da537815d9e8-osd--block--5368e13a--d21f--5912--91bd--da537815d9e8'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-m0F2nF-j9UB-mZfT-bRv8-xW1c-EVch-ZfDvHt', 'scsi-0QEMU_QEMU_HARDDISK_e03b171b-b829-41a0-a0b6-cc46d85f013d', 'scsi-SQEMU_QEMU_HARDDISK_e03b171b-b829-41a0-a0b6-cc46d85f013d'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_9fa3c153-0e2a-45ef-8e48-e0095eb937ad', 'scsi-SQEMU_QEMU_HARDDISK_9fa3c153-0e2a-45ef-8e48-e0095eb937ad'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-21-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part1', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part14', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part15', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part16', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--ede169e1--d974--55ed--a938--c303cb89cb5b-osd--block--ede169e1--d974--55ed--a938--c303cb89cb5b'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-HVhEyD-n8kM-PvHm-CUDD-tuRZ-igjC-Nzour2', 'scsi-0QEMU_QEMU_HARDDISK_f76ae8f0-65cc-4a62-b19c-23f8a40fbde6', 'scsi-SQEMU_QEMU_HARDDISK_f76ae8f0-65cc-4a62-b19c-23f8a40fbde6'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a-osd--block--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-YifJdB-JYWE-HhRV-mF1g-0k83-hzWG-ctJZWy', 'scsi-0QEMU_QEMU_HARDDISK_d7bb81ba-6491-4b5a-bb82-f0b49314eac0', 'scsi-SQEMU_QEMU_HARDDISK_d7bb81ba-6491-4b5a-bb82-f0b49314eac0'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_cdffe5f1-7f0d-460e-a1db-d6846d226144', 'scsi-SQEMU_QEMU_HARDDISK_cdffe5f1-7f0d-460e-a1db-d6846d226144'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-28-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5', 'scsi-SQEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part1', 'scsi-SQEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part14', 'scsi-SQEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part15', 'scsi-SQEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part16', 'scsi-SQEMU_QEMU_HARDDISK_64760a60-3a12-4fb7-90e5-3833dd4063b5-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f', 'scsi-SQEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part1', 'scsi-SQEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part14', 'scsi-SQEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part15', 'scsi-SQEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part16', 'scsi-SQEMU_QEMU_HARDDISK_7cc5c56e-2215-4807-a75d-aa809849b26f-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-23-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-10-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32', 'scsi-SQEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part1', 'scsi-SQEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part14', 'scsi-SQEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part15', 'scsi-SQEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part16', 'scsi-SQEMU_QEMU_HARDDISK_d6e05184-e73c-49c9-a2ec-7c822155aa32-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-19-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Saturday 21 March 2026  00:45:16 +0000 (0:00:01.861)       0:00:37.830 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Saturday 21 March 2026  00:45:17 +0000 (0:00:01.209)       0:00:39.040 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Saturday 21 March 2026  00:45:18 +0000 (0:00:00.771)       0:00:39.811 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Saturday 21 March 2026  00:45:19 +0000 (0:00:00.736)       0:00:40.547 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Saturday 21 March 2026  00:45:19 +0000 (0:00:00.535)       0:00:41.082 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Saturday 21 March 2026  00:45:20 +0000 (0:00:01.204)       0:00:42.287 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Saturday 21 March 2026  00:45:21 +0000 (0:00:01.089)       0:00:43.376 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Saturday 21 March 2026  00:45:25 +0000 (0:00:03.591)       0:00:46.968 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Saturday 21 March 2026  00:45:27 +0000 (0:00:01.868)       0:00:48.836 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Saturday 21 March 2026  00:45:29 +0000 (0:00:01.666)       0:00:50.503 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Saturday 21 March 2026  00:45:29 +0000 (0:00:00.312)       0:00:50.815 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Saturday 21 March 2026  00:45:29 +0000 (0:00:00.440)       0:00:51.256 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Saturday 21 March 2026  00:45:30 +0000 (0:00:00.376)       0:00:51.632 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Saturday 21 March 2026  00:45:31 +0000 (0:00:01.114)       0:00:52.747 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Saturday 21 March 2026  00:45:31 +0000 (0:00:00.435)       0:00:53.183 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Saturday 21 March 2026  00:45:32 +0000 (0:00:00.491)       0:00:53.675 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Saturday 21 March 2026  00:45:32 +0000 (0:00:00.405)       0:00:54.081 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Saturday 21 March 2026  00:45:32 +0000 (0:00:00.318)       0:00:54.400 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Saturday 21 March 2026  00:45:33 +0000 (0:00:00.712)       0:00:55.113 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Saturday 21 March 2026  00:45:34 +0000 (0:00:01.133)       0:00:56.246 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Saturday 21 March 2026  00:45:36 +0000 (0:00:02.014)       0:00:58.260 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Saturday 21 March 2026  00:45:37 +0000 (0:00:01.144)       0:00:59.404 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Saturday 21 March 2026  00:45:39 +0000 (0:00:01.251)       0:01:00.656 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Saturday 21 March 2026  00:45:40 +0000 (0:00:01.522)       0:01:02.179 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Saturday 21 March 2026  00:45:41 +0000 (0:00:00.984)       0:01:03.163 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Saturday 21 March 2026  00:45:42 +0000 (0:00:01.222)       0:01:04.386 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Saturday 21 March 2026  00:45:44 +0000 (0:00:01.779)       0:01:06.165 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Saturday 21 March 2026  00:45:45 +0000 (0:00:00.955)       0:01:07.121 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Saturday 21 March 2026  00:45:46 +0000 (0:00:00.983)       0:01:08.105 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Saturday 21 March 2026  00:45:47 +0000 (0:00:00.813)       0:01:08.919 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Saturday 21 March 2026  00:45:49 +0000 (0:00:02.087)       0:01:11.007 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Saturday 21 March 2026  00:45:51 +0000 (0:00:02.101)       0:01:13.108 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Saturday 21 March 2026  00:45:52 +0000 (0:00:01.278)       0:01:14.386 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Saturday 21 March 2026  00:45:53 +0000 (0:00:00.918)       0:01:15.304 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Saturday 21 March 2026  00:45:54 +0000 (0:00:00.874)       0:01:16.179 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Saturday 21 March 2026  00:45:55 +0000 (0:00:00.997)       0:01:17.177 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Saturday 21 March 2026  00:45:56 +0000 (0:00:01.029)       0:01:18.207 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Saturday 21 March 2026  00:45:57 +0000 (0:00:00.687)       0:01:18.895 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Saturday 21 March 2026  00:45:58 +0000 (0:00:01.004)       0:01:19.899 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Saturday 21 March 2026  00:45:59 +0000 (0:00:00.868)       0:01:20.768 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Saturday 21 March 2026  00:46:00 +0000 (0:00:00.906)       0:01:21.674 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Generate systemd ceph target file] ***************",
                                "Saturday 21 March 2026  00:46:01 +0000 (0:00:01.318)       0:01:22.993 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Enable ceph.target] ******************************",
                                "Saturday 21 March 2026  00:46:02 +0000 (0:00:01.346)       0:01:24.339 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include prerequisites.yml] ***********************",
                                "Saturday 21 March 2026  00:46:05 +0000 (0:00:02.188)       0:01:26.528 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Stop lvmetad] ************************************",
                                "Saturday 21 March 2026  00:46:06 +0000 (0:00:01.073)       0:01:27.602 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Disable and mask lvmetad service] ****************",
                                "Saturday 21 March 2026  00:46:06 +0000 (0:00:00.542)       0:01:28.145 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove ceph udev rules] **************************",
                                "Saturday 21 March 2026  00:46:07 +0000 (0:00:00.663)       0:01:28.808 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Ensure tmpfiles.d is present] ********************",
                                "Saturday 21 March 2026  00:46:08 +0000 (0:00:01.225)       0:01:30.034 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Restore certificates selinux context] ************",
                                "Saturday 21 March 2026  00:46:09 +0000 (0:00:00.974)       0:01:31.009 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Install python3 on osd nodes] ********************",
                                "Saturday 21 March 2026  00:46:10 +0000 (0:00:00.513)       0:01:31.523 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include registry.yml] ****************************",
                                "Saturday 21 March 2026  00:46:10 +0000 (0:00:00.668)       0:01:32.191 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include fetch_image.yml] *************************",
                                "Saturday 21 March 2026  00:46:11 +0000 (0:00:00.492)       0:01:32.684 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling Ceph container image] ********************",
                                "Saturday 21 March 2026  00:46:12 +0000 (0:00:01.094)       0:01:33.779 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***",
                                "Saturday 21 March 2026  00:47:10 +0000 (0:00:58.605)       0:02:32.384 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling node-exporter container image] ***********",
                                "Saturday 21 March 2026  00:47:11 +0000 (0:00:00.579)       0:02:32.963 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Export local ceph dev image] *********************",
                                "Saturday 21 March 2026  00:47:12 +0000 (0:00:00.640)       0:02:33.604 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Copy ceph dev image file] ************************",
                                "Saturday 21 March 2026  00:47:12 +0000 (0:00:00.150)       0:02:33.754 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Load ceph dev image] *****************************",
                                "Saturday 21 March 2026  00:47:12 +0000 (0:00:00.522)       0:02:34.277 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove tmp ceph dev image file] ******************",
                                "Saturday 21 March 2026  00:47:13 +0000 (0:00:00.742)       0:02:35.020 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Get ceph version] ********************************",
                                "Saturday 21 March 2026  00:47:14 +0000 (0:00:00.515)       0:02:35.535 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***",
                                "Saturday 21 March 2026  00:47:16 +0000 (0:00:02.715)       0:02:38.251 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include release.yml] *****************************",
                                "Saturday 21 March 2026  00:47:17 +0000 (0:00:00.662)       0:02:38.913 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release jewel] *********************",
                                "Saturday 21 March 2026  00:47:18 +0000 (0:00:01.095)       0:02:40.009 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release kraken] ********************",
                                "Saturday 21 March 2026  00:47:19 +0000 (0:00:00.533)       0:02:40.542 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release luminous] ******************",
                                "Saturday 21 March 2026  00:47:19 +0000 (0:00:00.784)       0:02:41.326 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release mimic] *********************",
                                "Saturday 21 March 2026  00:47:20 +0000 (0:00:00.559)       0:02:41.886 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release nautilus] ******************",
                                "Saturday 21 March 2026  00:47:21 +0000 (0:00:00.741)       0:02:42.628 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release octopus] *******************",
                                "Saturday 21 March 2026  00:47:21 +0000 (0:00:00.517)       0:02:43.146 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release pacific] *******************",
                                "Saturday 21 March 2026  00:47:22 +0000 (0:00:00.685)       0:02:43.831 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release quincy] ********************",
                                "Saturday 21 March 2026  00:47:22 +0000 (0:00:00.534)       0:02:44.366 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release reef] **********************",
                                "Saturday 21 March 2026  00:47:23 +0000 (0:00:00.683)       0:02:45.050 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************",
                                "Saturday 21 March 2026  00:47:25 +0000 (0:00:01.434)       0:02:46.485 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph initial directories] ***************************",
                                "Saturday 21 March 2026  00:47:26 +0000 (0:00:01.148)       0:02:47.633 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/log/ceph)\u001b[0m",
                                "",
                                "TASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************",
                                "Saturday 21 March 2026  00:47:32 +0000 (0:00:06.694)       0:02:54.328 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-config : Create rados gateway instance directories] *****************",
                                "Saturday 21 March 2026  00:47:34 +0000 (0:00:01.234)       0:02:55.562 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate environment file] *********************************",
                                "Saturday 21 March 2026  00:47:34 +0000 (0:00:00.628)       0:02:56.191 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Reset num_osds] ********************************************",
                                "Saturday 21 March 2026  00:47:35 +0000 (0:00:01.177)       0:02:57.369 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Count number of osds for lvm scenario] *********************",
                                "Saturday 21 March 2026  00:47:36 +0000 (0:00:00.562)       0:02:57.931 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Look up for ceph-volume rejected devices] ******************",
                                "Saturday 21 March 2026  00:47:37 +0000 (0:00:00.843)       0:02:58.775 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact rejected_devices] *********************************",
                                "Saturday 21 March 2026  00:47:38 +0000 (0:00:00.666)       0:02:59.442 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _devices] *****************************************",
                                "Saturday 21 March 2026  00:47:38 +0000 (0:00:00.846)       0:03:00.288 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***",
                                "Saturday 21 March 2026  00:47:40 +0000 (0:00:01.439)       0:03:01.728 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***",
                                "Saturday 21 March 2026  00:47:40 +0000 (0:00:00.644)       0:03:02.372 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***",
                                "Saturday 21 March 2026  00:47:41 +0000 (0:00:00.706)       0:03:03.078 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***",
                                "Saturday 21 March 2026  00:47:42 +0000 (0:00:00.789)       0:03:03.868 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds (add existing osds)] *********************",
                                "Saturday 21 March 2026  00:47:45 +0000 (0:00:03.449)       0:03:07.317 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _osd_memory_target] *******************************",
                                "Saturday 21 March 2026  00:47:46 +0000 (0:00:00.543)       0:03:07.861 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set osd_memory_target to cluster host config] **************",
                                "Saturday 21 March 2026  00:47:47 +0000 (0:00:00.799)       0:03:08.661 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Render rgw configs] ****************************************",
                                "Saturday 21 March 2026  00:47:47 +0000 (0:00:00.582)       0:03:09.243 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set config to cluster] *************************************",
                                "Saturday 21 March 2026  00:47:48 +0000 (0:00:00.708)       0:03:09.952 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.13:8081'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.14:8081'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.15:8081'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set rgw configs to file] ***********************************",
                                "Saturday 21 March 2026  00:47:49 +0000 (0:00:00.580)       0:03:10.532 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph conf directory] ********************************",
                                "Saturday 21 March 2026  00:47:50 +0000 (0:00:00.919)       0:03:11.452 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Saturday 21 March 2026  00:47:50 +0000 (0:00:00.728)       0:03:12.181 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Saturday 21 March 2026  00:47:51 +0000 (0:00:01.079)       0:03:13.260 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Saturday 21 March 2026  00:47:52 +0000 (0:00:00.741)       0:03:14.002 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Saturday 21 March 2026  00:47:53 +0000 (0:00:00.982)       0:03:14.985 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Saturday 21 March 2026  00:47:54 +0000 (0:00:00.695)       0:03:15.680 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Saturday 21 March 2026  00:47:54 +0000 (0:00:00.540)       0:03:16.221 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Saturday 21 March 2026  00:47:55 +0000 (0:00:00.582)       0:03:16.803 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Saturday 21 March 2026  00:47:56 +0000 (0:00:00.700)       0:03:17.503 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Saturday 21 March 2026  00:47:56 +0000 (0:00:00.490)       0:03:17.993 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=0) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate Ceph file] ****************************************",
                                "Saturday 21 March 2026  00:47:57 +0000 (0:00:01.344)       0:03:19.338 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Saturday 21 March 2026  00:48:00 +0000 (0:00:02.535)       0:03:21.873 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mons handler] **********************************",
                                "Saturday 21 March 2026  00:48:01 +0000 (0:00:01.381)       0:03:23.254 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mons.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mon_handler_called before restart] ********",
                                "Saturday 21 March 2026  00:48:02 +0000 (0:00:00.778)       0:03:24.033 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mon restart script] ***********************",
                                "Saturday 21 March 2026  00:48:02 +0000 (0:00:00.331)       0:03:24.364 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mon daemon(s)] ********************",
                                "Saturday 21 March 2026  00:48:04 +0000 (0:00:01.247)       0:03:25.612 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mon_handler_called after restart] *********",
                                "Saturday 21 March 2026  00:48:04 +0000 (0:00:00.787)       0:03:26.399 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Osds handler] **********************************",
                                "Saturday 21 March 2026  00:48:05 +0000 (0:00:00.281)       0:03:26.680 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set_fact trigger_restart] **********************",
                                "Saturday 21 March 2026  00:48:06 +0000 (0:00:01.206)       0:03:27.887 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _osd_handler_called before restart] ********",
                                "Saturday 21 March 2026  00:48:06 +0000 (0:00:00.405)       0:03:28.292 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Unset noup flag] *******************************",
                                "Saturday 21 March 2026  00:48:07 +0000 (0:00:00.674)       0:03:28.967 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy osd restart script] ***********************",
                                "Saturday 21 March 2026  00:48:07 +0000 (0:00:00.218)       0:03:29.186 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Get pool list] *********************************",
                                "Saturday 21 March 2026  00:48:08 +0000 (0:00:00.286)       0:03:29.473 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Get balancer module status] ********************",
                                "Saturday 21 March 2026  00:48:08 +0000 (0:00:00.258)       0:03:29.731 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set_fact pools_pgautoscaler_mode] **************",
                                "Saturday 21 March 2026  00:48:08 +0000 (0:00:00.174)       0:03:29.906 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Disable balancer] ******************************",
                                "Saturday 21 March 2026  00:48:08 +0000 (0:00:00.105)       0:03:30.012 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Disable pg autoscale on pools] *****************",
                                "Saturday 21 March 2026  00:48:08 +0000 (0:00:00.220)       0:03:30.233 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph osds daemon(s)] *******************",
                                "Saturday 21 March 2026  00:48:09 +0000 (0:00:00.216)       0:03:30.449 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _osd_handler_called after restart] *********",
                                "Saturday 21 March 2026  00:48:09 +0000 (0:00:00.550)       0:03:31.000 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Re-enable pg autoscale on pools] ***************",
                                "Saturday 21 March 2026  00:48:10 +0000 (0:00:00.441)       0:03:31.441 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Re-enable balancer] ****************************",
                                "Saturday 21 March 2026  00:48:10 +0000 (0:00:00.225)       0:03:31.667 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mdss handler] **********************************",
                                "Saturday 21 March 2026  00:48:10 +0000 (0:00:00.232)       0:03:31.900 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mdss.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mds_handler_called before restart] ********",
                                "Saturday 21 March 2026  00:48:11 +0000 (0:00:00.813)       0:03:32.713 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mds restart script] ***********************",
                                "Saturday 21 March 2026  00:48:11 +0000 (0:00:00.440)       0:03:33.154 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mds daemon(s)] ********************",
                                "Saturday 21 March 2026  00:48:12 +0000 (0:00:00.994)       0:03:34.148 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mds_handler_called after restart] *********",
                                "Saturday 21 March 2026  00:48:13 +0000 (0:00:00.536)       0:03:34.685 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rgws handler] **********************************",
                                "Saturday 21 March 2026  00:48:13 +0000 (0:00:00.258)       0:03:34.943 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_rgws.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _rgw_handler_called before restart] ********",
                                "Saturday 21 March 2026  00:48:14 +0000 (0:00:00.905)       0:03:35.849 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy rgw restart script] ***********************",
                                "Saturday 21 March 2026  00:48:14 +0000 (0:00:00.316)       0:03:36.165 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph rgw daemon(s)] ********************",
                                "Saturday 21 March 2026  00:48:15 +0000 (0:00:01.221)       0:03:37.387 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _rgw_handler_called after restart] *********",
                                "Saturday 21 March 2026  00:48:16 +0000 (0:00:00.554)       0:03:37.941 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rbdmirrors handler] ****************************",
                                "Saturday 21 March 2026  00:48:16 +0000 (0:00:00.349)       0:03:38.291 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mgrs handler] **********************************",
                                "Saturday 21 March 2026  00:48:17 +0000 (0:00:00.605)       0:03:38.896 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mgrs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mgr_handler_called before restart] ********",
                                "Saturday 21 March 2026  00:48:18 +0000 (0:00:00.977)       0:03:39.873 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mgr restart script] ***********************",
                                "Saturday 21 March 2026  00:48:18 +0000 (0:00:00.277)       0:03:40.151 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mgr daemon(s)] ********************",
                                "Saturday 21 March 2026  00:48:20 +0000 (0:00:01.376)       0:03:41.527 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mgr_handler_called after restart] *********",
                                "Saturday 21 March 2026  00:48:20 +0000 (0:00:00.528)       0:03:42.056 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-mon] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Saturday 21 March 2026  00:48:21 +0000 (0:00:00.490)       0:03:42.546 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Saturday 21 March 2026  00:48:21 +0000 (0:00:00.601)       0:03:43.148 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Saturday 21 March 2026  00:48:22 +0000 (0:00:00.494)       0:03:43.642 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Saturday 21 March 2026  00:48:22 +0000 (0:00:00.621)       0:03:44.263 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Saturday 21 March 2026  00:48:23 +0000 (0:00:00.285)       0:03:44.549 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Saturday 21 March 2026  00:48:23 +0000 (0:00:00.452)       0:03:45.002 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Saturday 21 March 2026  00:48:23 +0000 (0:00:00.307)       0:03:45.310 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Saturday 21 March 2026  00:48:24 +0000 (0:00:00.615)       0:03:45.925 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Saturday 21 March 2026  00:48:24 +0000 (0:00:00.284)       0:03:46.210 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Saturday 21 March 2026  00:48:25 +0000 (0:00:00.495)       0:03:46.705 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Saturday 21 March 2026  00:48:26 +0000 (0:00:00.795)       0:03:47.501 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Saturday 21 March 2026  00:48:26 +0000 (0:00:00.712)       0:03:48.213 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Saturday 21 March 2026  00:48:27 +0000 (0:00:00.266)       0:03:48.480 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Saturday 21 March 2026  00:48:27 +0000 (0:00:00.441)       0:03:48.922 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Saturday 21 March 2026  00:48:27 +0000 (0:00:00.320)       0:03:49.243 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Saturday 21 March 2026  00:48:28 +0000 (0:00:00.253)       0:03:49.496 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Saturday 21 March 2026  00:48:28 +0000 (0:00:00.257)       0:03:49.754 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Saturday 21 March 2026  00:48:28 +0000 (0:00:00.448)       0:03:50.203 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Saturday 21 March 2026  00:48:29 +0000 (0:00:00.261)       0:03:50.464 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Saturday 21 March 2026  00:48:29 +0000 (0:00:00.292)       0:03:50.757 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Saturday 21 March 2026  00:48:29 +0000 (0:00:00.296)       0:03:51.054 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact container_exec_cmd] **********************************",
                                "Saturday 21 March 2026  00:48:30 +0000 (0:00:00.749)       0:03:51.803 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include deploy_monitors.yml] **********************************",
                                "Saturday 21 March 2026  00:48:30 +0000 (0:00:00.287)       0:03:52.091 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/deploy_monitors.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mon : Check if monitor initial keyring already exists] **************",
                                "Saturday 21 March 2026  00:48:31 +0000 (0:00:00.478)       0:03:52.569 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate monitor initial keyring] *****************************",
                                "Saturday 21 March 2026  00:48:31 +0000 (0:00:00.375)       0:03:52.944 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact _initial_mon_key_success] ****************************",
                                "Saturday 21 March 2026  00:48:32 +0000 (0:00:01.037)       0:03:53.982 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Get initial keyring when it already exists] *******************",
                                "Saturday 21 March 2026  00:48:33 +0000 (0:00:00.439)       0:03:54.421 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Create monitor initial keyring] *******************************",
                                "Saturday 21 March 2026  00:48:33 +0000 (0:00:00.358)       0:03:54.780 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Copy the initial key in /etc/ceph (for containers)] ***********",
                                "Saturday 21 March 2026  00:48:34 +0000 (0:00:01.014)       0:03:55.794 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Create monitor directory] *************************************",
                                "Saturday 21 March 2026  00:48:35 +0000 (0:00:01.002)       0:03:56.797 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Recursively fix ownership of monitor directory] ***************",
                                "Saturday 21 March 2026  00:48:36 +0000 (0:00:00.713)       0:03:57.510 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Create admin keyring] *****************************************",
                                "Saturday 21 March 2026  00:48:36 +0000 (0:00:00.680)       0:03:58.191 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Slurp admin keyring] ******************************************",
                                "Saturday 21 March 2026  00:48:37 +0000 (0:00:01.156)       0:03:59.347 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Copy admin keyring over to mons] ******************************",
                                "Saturday 21 March 2026  00:48:38 +0000 (0:00:00.738)       0:04:00.086 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0 -> {{ item }}]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> {{ item }}]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> {{ item }}]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Import admin keyring into mon keyring] ************************",
                                "Saturday 21 March 2026  00:48:42 +0000 (0:00:04.013)       0:04:04.099 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact ceph-mon container command] **************************",
                                "Saturday 21 March 2026  00:48:44 +0000 (0:00:01.350)       0:04:05.450 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact monmaptool container command] ************************",
                                "Saturday 21 March 2026  00:48:44 +0000 (0:00:00.417)       0:04:05.867 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate initial monmap] **************************************",
                                "Saturday 21 March 2026  00:48:44 +0000 (0:00:00.439)       0:04:06.307 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Ceph monitor mkfs with keyring] *******************************",
                                "Saturday 21 March 2026  00:48:47 +0000 (0:00:02.405)       0:04:08.712 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Ceph monitor mkfs without keyring] ****************************",
                                "Saturday 21 March 2026  00:48:49 +0000 (0:00:01.738)       0:04:10.451 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include start_monitor.yml] ************************************",
                                "Saturday 21 March 2026  00:48:49 +0000 (0:00:00.279)       0:04:10.730 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/start_monitor.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mon : Ensure systemd service override directory exists] *************",
                                "Saturday 21 March 2026  00:48:49 +0000 (0:00:00.421)       0:04:11.152 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Add ceph-mon systemd service overrides] ***********************",
                                "Saturday 21 March 2026  00:48:50 +0000 (0:00:00.968)       0:04:12.121 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include_tasks systemd.yml] ************************************",
                                "Saturday 21 March 2026  00:48:51 +0000 (0:00:00.415)       0:04:12.536 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/systemd.yml for testbed-node-0, testbed-node-2, testbed-node-1\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate systemd unit file for mon container] *****************",
                                "Saturday 21 March 2026  00:48:52 +0000 (0:00:00.922)       0:04:13.458 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate systemd ceph-mon target file] ************************",
                                "Saturday 21 March 2026  00:48:53 +0000 (0:00:01.925)       0:04:15.383 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Enable ceph-mon.target] ***************************************",
                                "Saturday 21 March 2026  00:48:55 +0000 (0:00:01.481)       0:04:16.865 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Start the monitor service] ************************************",
                                "Saturday 21 March 2026  00:48:57 +0000 (0:00:01.710)       0:04:18.575 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include_tasks ceph_keys.yml] **********************************",
                                "Saturday 21 March 2026  00:48:59 +0000 (0:00:01.942)       0:04:20.517 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/ceph_keys.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mon : Waiting for the monitor(s) to form the quorum...] *************",
                                "Saturday 21 March 2026  00:48:59 +0000 (0:00:00.651)       0:04:21.169 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for the monitor(s) to form the quorum... (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Fetch ceph initial keys] **************************************",
                                "Saturday 21 March 2026  00:49:21 +0000 (0:00:21.519)       0:04:42.689 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include secure_cluster.yml] ***********************************",
                                "Saturday 21 March 2026  00:49:30 +0000 (0:00:08.912)       0:04:51.601 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set cluster configs] ******************************************",
                                "Saturday 21 March 2026  00:49:30 +0000 (0:00:00.255)       0:04:51.857 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__8519e6c23c9c83674cbc6ea19a79c2df830b7b57'}}, {'key': 'public_network', 'value': '192.168.16.0/20'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__8519e6c23c9c83674cbc6ea19a79c2df830b7b57'}}, {'key': 'cluster_network', 'value': '192.168.16.0/20'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__8519e6c23c9c83674cbc6ea19a79c2df830b7b57'}}, {'key': 'osd_pool_default_crush_rule', 'value': -1}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__8519e6c23c9c83674cbc6ea19a79c2df830b7b57'}}, {'key': 'ms_bind_ipv6', 'value': 'False'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__8519e6c23c9c83674cbc6ea19a79c2df830b7b57'}}, {'key': 'ms_bind_ipv4', 'value': 'True'}])\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__8519e6c23c9c83674cbc6ea19a79c2df830b7b57'}}, {'key': 'osd_crush_chooseleaf_type', 'value': '__omit_place_holder__8519e6c23c9c83674cbc6ea19a79c2df830b7b57'}]) \u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Saturday 21 March 2026  00:49:45 +0000 (0:00:14.955)       0:05:06.812 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mons handler] **********************************",
                                "Saturday 21 March 2026  00:49:45 +0000 (0:00:00.322)       0:05:07.135 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mons.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mon_handler_called before restart] ********",
                                "Saturday 21 March 2026  00:49:46 +0000 (0:00:00.477)       0:05:07.613 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mon restart script] ***********************",
                                "Saturday 21 March 2026  00:49:46 +0000 (0:00:00.424)       0:05:08.037 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mon daemon(s)] ********************",
                                "Saturday 21 March 2026  00:49:46 +0000 (0:00:00.277)       0:05:08.314 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mon_handler_called after restart] *********",
                                "Saturday 21 March 2026  00:49:47 +0000 (0:00:00.537)       0:05:08.851 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-mgr] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Saturday 21 March 2026  00:49:48 +0000 (0:00:00.632)       0:05:09.484 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Saturday 21 March 2026  00:49:48 +0000 (0:00:00.431)       0:05:09.915 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Saturday 21 March 2026  00:49:48 +0000 (0:00:00.464)       0:05:10.379 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Saturday 21 March 2026  00:49:49 +0000 (0:00:00.866)       0:05:11.246 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Saturday 21 March 2026  00:49:50 +0000 (0:00:00.267)       0:05:11.513 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Saturday 21 March 2026  00:49:50 +0000 (0:00:00.304)       0:05:11.817 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Saturday 21 March 2026  00:49:50 +0000 (0:00:00.298)       0:05:12.116 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Saturday 21 March 2026  00:49:51 +0000 (0:00:00.971)       0:05:13.087 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Saturday 21 March 2026  00:49:51 +0000 (0:00:00.279)       0:05:13.367 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Saturday 21 March 2026  00:49:52 +0000 (0:00:00.262)       0:05:13.630 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Saturday 21 March 2026  00:49:52 +0000 (0:00:00.691)       0:05:14.321 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Saturday 21 March 2026  00:49:53 +0000 (0:00:00.860)       0:05:15.181 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Saturday 21 March 2026  00:49:54 +0000 (0:00:00.274)       0:05:15.456 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Saturday 21 March 2026  00:49:54 +0000 (0:00:00.275)       0:05:15.731 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Saturday 21 March 2026  00:49:54 +0000 (0:00:00.246)       0:05:15.978 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Saturday 21 March 2026  00:49:54 +0000 (0:00:00.403)       0:05:16.381 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Saturday 21 March 2026  00:49:55 +0000 (0:00:00.256)       0:05:16.638 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Saturday 21 March 2026  00:49:55 +0000 (0:00:00.250)       0:05:16.888 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Saturday 21 March 2026  00:49:55 +0000 (0:00:00.272)       0:05:17.161 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Saturday 21 March 2026  00:49:55 +0000 (0:00:00.258)       0:05:17.420 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Saturday 21 March 2026  00:49:56 +0000 (0:00:00.490)       0:05:17.910 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set_fact container_exec_cmd] **********************************",
                                "Saturday 21 March 2026  00:49:56 +0000 (0:00:00.445)       0:05:18.356 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include common.yml] *******************************************",
                                "Saturday 21 March 2026  00:49:57 +0000 (0:00:00.700)       0:05:19.056 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/common.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Create mgr directory] *****************************************",
                                "Saturday 21 March 2026  00:49:58 +0000 (0:00:00.609)       0:05:19.666 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Fetch ceph mgr keyring] ***************************************",
                                "Saturday 21 March 2026  00:49:58 +0000 (0:00:00.658)       0:05:20.325 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Create ceph mgr keyring(s) on a mon node] *********************",
                                "Saturday 21 March 2026  00:49:59 +0000 (0:00:00.264)       0:05:20.589 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0 -> {{ groups[mon_group_name][0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set_fact _mgr_keys] *******************************************",
                                "Saturday 21 March 2026  00:50:09 +0000 (0:00:10.782)       0:05:31.372 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Get keys from monitors] ***************************************",
                                "Saturday 21 March 2026  00:50:10 +0000 (0:00:00.686)       0:05:32.058 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Copy ceph key(s) if needed] ***********************************",
                                "Saturday 21 March 2026  00:50:13 +0000 (0:00:02.652)       0:05:34.711 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=None)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set mgr key permissions] **************************************",
                                "Saturday 21 March 2026  00:50:14 +0000 (0:00:01.345)       0:05:36.056 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Append dashboard modules to ceph_mgr_modules] *****************",
                                "Saturday 21 March 2026  00:50:15 +0000 (0:00:00.788)       0:05:36.844 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include pre_requisite.yml] ************************************",
                                "Saturday 21 March 2026  00:50:15 +0000 (0:00:00.547)       0:05:37.392 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include start_mgr.yml] ****************************************",
                                "Saturday 21 March 2026  00:50:16 +0000 (0:00:00.278)       0:05:37.670 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/start_mgr.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Ensure systemd service override directory exists] *************",
                                "Saturday 21 March 2026  00:50:16 +0000 (0:00:00.454)       0:05:38.124 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Add ceph-mgr systemd service overrides] ***********************",
                                "Saturday 21 March 2026  00:50:16 +0000 (0:00:00.275)       0:05:38.399 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include_tasks systemd.yml] ************************************",
                                "Saturday 21 March 2026  00:50:17 +0000 (0:00:00.473)       0:05:38.873 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/systemd.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Generate systemd unit file] ***********************************",
                                "Saturday 21 March 2026  00:50:17 +0000 (0:00:00.433)       0:05:39.306 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Generate systemd ceph-mgr target file] ************************",
                                "Saturday 21 March 2026  00:50:19 +0000 (0:00:01.191)       0:05:40.498 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Enable ceph-mgr.target] ***************************************",
                                "Saturday 21 March 2026  00:50:20 +0000 (0:00:01.392)       0:05:41.890 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Systemd start mgr] ********************************************",
                                "Saturday 21 March 2026  00:50:22 +0000 (0:00:01.970)       0:05:43.860 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include mgr_modules.yml] **************************************",
                                "Saturday 21 March 2026  00:50:24 +0000 (0:00:01.986)       0:05:45.847 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/mgr_modules.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Wait for all mgr to be up] ************************************",
                                "Saturday 21 March 2026  00:50:24 +0000 (0:00:00.389)       0:05:46.237 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (30 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (29 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (28 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (27 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (26 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (25 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Get enabled modules from ceph-mgr] ****************************",
                                "Saturday 21 March 2026  00:51:01 +0000 (0:00:36.711)       0:06:22.949 ******** ",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set _ceph_mgr_modules fact (convert _ceph_mgr_modules.stdout to a dict)] ***",
                                "Saturday 21 March 2026  00:51:02 +0000 (0:00:01.236)       0:06:24.185 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set _disabled_ceph_mgr_modules fact] **************************",
                                "Saturday 21 March 2026  00:51:03 +0000 (0:00:00.300)       0:06:24.486 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Disable ceph mgr enabled modules] *****************************",
                                "Saturday 21 March 2026  00:51:03 +0000 (0:00:00.115)       0:06:24.601 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=iostat)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=nfs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=restful)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Add modules to ceph-mgr] **************************************",
                                "Saturday 21 March 2026  00:51:09 +0000 (0:00:06.385)       0:06:30.986 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=balancer) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=dashboard)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=prometheus)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=status) \u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Saturday 21 March 2026  00:51:14 +0000 (0:00:04.811)       0:06:35.798 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mgrs handler] **********************************",
                                "Saturday 21 March 2026  00:51:15 +0000 (0:00:01.012)       0:06:36.811 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mgrs.yml for testbed-node-1, testbed-node-0, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mgr_handler_called before restart] ********",
                                "Saturday 21 March 2026  00:51:16 +0000 (0:00:00.614)       0:06:37.425 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mgr restart script] ***********************",
                                "Saturday 21 March 2026  00:51:16 +0000 (0:00:00.320)       0:06:37.745 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mgr daemon(s)] ********************",
                                "Saturday 21 March 2026  00:51:17 +0000 (0:00:01.368)       0:06:39.114 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mgr_handler_called after restart] *********",
                                "Saturday 21 March 2026  00:51:18 +0000 (0:00:00.575)       0:06:39.689 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-osd] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Saturday 21 March 2026  00:51:18 +0000 (0:00:00.491)       0:06:40.181 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Saturday 21 March 2026  00:51:19 +0000 (0:00:00.574)       0:06:40.755 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-5, testbed-node-4\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Saturday 21 March 2026  00:51:19 +0000 (0:00:00.489)       0:06:41.245 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Saturday 21 March 2026  00:51:20 +0000 (0:00:00.294)       0:06:41.540 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Saturday 21 March 2026  00:51:20 +0000 (0:00:00.800)       0:06:42.341 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Saturday 21 March 2026  00:51:21 +0000 (0:00:00.631)       0:06:42.973 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Saturday 21 March 2026  00:51:22 +0000 (0:00:00.727)       0:06:43.701 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Saturday 21 March 2026  00:51:22 +0000 (0:00:00.252)       0:06:43.953 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Saturday 21 March 2026  00:51:22 +0000 (0:00:00.406)       0:06:44.359 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Saturday 21 March 2026  00:51:23 +0000 (0:00:00.268)       0:06:44.628 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Saturday 21 March 2026  00:51:23 +0000 (0:00:00.626)       0:06:45.254 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Saturday 21 March 2026  00:51:24 +0000 (0:00:00.588)       0:06:45.843 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Saturday 21 March 2026  00:51:24 +0000 (0:00:00.469)       0:06:46.312 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Saturday 21 March 2026  00:51:25 +0000 (0:00:00.249)       0:06:46.562 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Saturday 21 March 2026  00:51:25 +0000 (0:00:00.285)       0:06:46.848 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Saturday 21 March 2026  00:51:25 +0000 (0:00:00.280)       0:06:47.128 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Saturday 21 March 2026  00:51:26 +0000 (0:00:00.487)       0:06:47.616 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Saturday 21 March 2026  00:51:26 +0000 (0:00:00.283)       0:06:47.899 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Saturday 21 March 2026  00:51:26 +0000 (0:00:00.246)       0:06:48.146 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Saturday 21 March 2026  00:51:26 +0000 (0:00:00.255)       0:06:48.402 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Saturday 21 March 2026  00:51:27 +0000 (0:00:00.525)       0:06:48.928 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set_fact add_osd] *********************************************",
                                "Saturday 21 March 2026  00:51:27 +0000 (0:00:00.458)       0:06:49.386 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set_fact container_exec_cmd] **********************************",
                                "Saturday 21 March 2026  00:51:28 +0000 (0:00:00.278)       0:06:49.665 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks system_tuning.yml] ******************************",
                                "Saturday 21 March 2026  00:51:28 +0000 (0:00:00.714)       0:06:50.380 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/system_tuning.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Create tmpfiles.d directory] **********************************",
                                "Saturday 21 March 2026  00:51:29 +0000 (0:00:00.644)       0:06:51.024 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Disable transparent hugepage] *********************************",
                                "Saturday 21 March 2026  00:51:29 +0000 (0:00:00.241)       0:06:51.266 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Get default vm.min_free_kbytes] *******************************",
                                "Saturday 21 March 2026  00:51:30 +0000 (0:00:00.272)       0:06:51.538 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set_fact vm_min_free_kbytes] **********************************",
                                "Saturday 21 March 2026  00:51:30 +0000 (0:00:00.858)       0:06:52.397 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Apply operating system tuning] ********************************",
                                "Saturday 21 March 2026  00:51:31 +0000 (0:00:00.351)       0:06:52.748 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m",
                                "",
                                "TASK [ceph-osd : Install dependencies] *****************************************",
                                "Saturday 21 March 2026  00:51:35 +0000 (0:00:03.956)       0:06:56.704 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks common.yml] *************************************",
                                "Saturday 21 March 2026  00:51:35 +0000 (0:00:00.253)       0:06:56.958 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Create bootstrap-osd and osd directories] *********************",
                                "Saturday 21 March 2026  00:51:36 +0000 (0:00:00.612)       0:06:57.570 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/osd/)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Get keys from monitors] ***************************************",
                                "Saturday 21 March 2026  00:51:37 +0000 (0:00:00.955)       0:06:58.526 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Copy ceph key(s) if needed] ***********************************",
                                "Saturday 21 March 2026  00:51:39 +0000 (0:00:02.197)       0:07:00.724 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set noup flag] ************************************************",
                                "Saturday 21 March 2026  00:51:40 +0000 (0:00:01.456)       0:07:02.180 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks scenarios/lvm.yml] ******************************",
                                "Saturday 21 March 2026  00:51:42 +0000 (0:00:02.154)       0:07:04.334 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/scenarios/lvm.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Use ceph-volume to create osds] *******************************",
                                "Saturday 21 March 2026  00:51:43 +0000 (0:00:00.457)       0:07:04.792 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-64dd100a-6bcc-5763-b557-0732c79bfd55', 'data_vg': 'ceph-64dd100a-6bcc-5763-b557-0732c79bfd55'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-ede169e1-d974-55ed-a938-c303cb89cb5b', 'data_vg': 'ceph-ede169e1-d974-55ed-a938-c303cb89cb5b'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-ed04aa0b-4f5a-523b-a5d2-05072a1b260d', 'data_vg': 'ceph-ed04aa0b-4f5a-523b-a5d2-05072a1b260d'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a', 'data_vg': 'ceph-23a81e86-5308-55a4-b4d4-ad2cd0b09f4a'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-5368e13a-d21f-5912-91bd-da537815d9e8', 'data_vg': 'ceph-5368e13a-d21f-5912-91bd-da537815d9e8'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-45b09425-beda-50b4-a83b-1fc50b5adca4', 'data_vg': 'ceph-45b09425-beda-50b4-a83b-1fc50b5adca4'})\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks scenarios/lvm-batch.yml] ************************",
                                "Saturday 21 March 2026  00:52:27 +0000 (0:00:44.172)       0:07:48.964 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks start_osds.yml] *********************************",
                                "Saturday 21 March 2026  00:52:28 +0000 (0:00:00.470)       0:07:49.434 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/start_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Get osd ids] **************************************************",
                                "Saturday 21 March 2026  00:52:28 +0000 (0:00:00.438)       0:07:49.873 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Collect osd ids] **********************************************",
                                "Saturday 21 March 2026  00:52:29 +0000 (0:00:00.662)       0:07:50.535 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks systemd.yml] ************************************",
                                "Saturday 21 March 2026  00:52:32 +0000 (0:00:02.920)       0:07:53.455 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Generate systemd unit file] ***********************************",
                                "Saturday 21 March 2026  00:52:32 +0000 (0:00:00.470)       0:07:53.926 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Generate systemd ceph-osd target file] ************************",
                                "Saturday 21 March 2026  00:52:33 +0000 (0:00:01.227)       0:07:55.153 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Enable ceph-osd.target] ***************************************",
                                "Saturday 21 March 2026  00:52:35 +0000 (0:00:01.382)       0:07:56.535 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Ensure systemd service override directory exists] *************",
                                "Saturday 21 March 2026  00:52:36 +0000 (0:00:01.865)       0:07:58.401 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Add ceph-osd systemd service overrides] ***********************",
                                "Saturday 21 March 2026  00:52:37 +0000 (0:00:00.322)       0:07:58.723 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Ensure /var/lib/ceph/osd/<cluster>-<item> is present] *********",
                                "Saturday 21 March 2026  00:52:37 +0000 (0:00:00.285)       0:07:59.009 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=2)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Write run file in /var/lib/ceph/osd/xxxx/run] *****************",
                                "Saturday 21 March 2026  00:52:38 +0000 (0:00:01.284)       0:08:00.294 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=1)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=4)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=5)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=3)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=2)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Systemd start osd] ********************************************",
                                "Saturday 21 March 2026  00:52:41 +0000 (0:00:02.153)       0:08:02.448 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=1)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=4)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=3)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=5)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Unset noup flag] **********************************************",
                                "Saturday 21 March 2026  00:52:44 +0000 (0:00:03.861)       0:08:06.309 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Wait for all osd to be up] ************************************",
                                "Saturday 21 March 2026  00:52:47 +0000 (0:00:03.044)       0:08:09.354 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-5 -> testbed-node-0]: Wait for all osd to be up (60 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include crush_rules.yml] **************************************",
                                "Saturday 21 March 2026  00:53:00 +0000 (0:00:12.863)       0:08:22.217 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Saturday 21 March 2026  00:53:01 +0000 (0:00:00.738)       0:08:22.956 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Osds handler] **********************************",
                                "Saturday 21 March 2026  00:53:02 +0000 (0:00:00.475)       0:08:23.431 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set_fact trigger_restart] **********************",
                                "Saturday 21 March 2026  00:53:02 +0000 (0:00:00.447)       0:08:23.879 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _osd_handler_called before restart] ********",
                                "Saturday 21 March 2026  00:53:02 +0000 (0:00:00.354)       0:08:24.233 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Unset noup flag] *******************************",
                                "Saturday 21 March 2026  00:53:03 +0000 (0:00:00.289)       0:08:24.523 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy osd restart script] ***********************",
                                "Saturday 21 March 2026  00:53:03 +0000 (0:00:00.625)       0:08:25.148 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Get pool list] *********************************",
                                "Saturday 21 March 2026  00:53:03 +0000 (0:00:00.259)       0:08:25.408 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Get balancer module status] ********************",
                                "Saturday 21 March 2026  00:53:04 +0000 (0:00:00.206)       0:08:25.615 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set_fact pools_pgautoscaler_mode] **************",
                                "Saturday 21 March 2026  00:53:04 +0000 (0:00:00.217)       0:08:25.832 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Disable balancer] ******************************",
                                "Saturday 21 March 2026  00:53:04 +0000 (0:00:00.108)       0:08:25.941 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Disable pg autoscale on pools] *****************",
                                "Saturday 21 March 2026  00:53:04 +0000 (0:00:00.200)       0:08:26.141 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph osds daemon(s)] *******************",
                                "Saturday 21 March 2026  00:53:04 +0000 (0:00:00.191)       0:08:26.333 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _osd_handler_called after restart] *********",
                                "Saturday 21 March 2026  00:53:05 +0000 (0:00:00.364)       0:08:26.698 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Re-enable pg autoscale on pools] ***************",
                                "Saturday 21 March 2026  00:53:05 +0000 (0:00:00.299)       0:08:26.997 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Re-enable balancer] ****************************",
                                "Saturday 21 March 2026  00:53:06 +0000 (0:00:00.660)       0:08:27.658 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-crash] ***************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Saturday 21 March 2026  00:53:06 +0000 (0:00:00.601)       0:08:28.259 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Saturday 21 March 2026  00:53:07 +0000 (0:00:01.061)       0:08:29.321 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Saturday 21 March 2026  00:53:09 +0000 (0:00:01.185)       0:08:30.507 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Saturday 21 March 2026  00:53:10 +0000 (0:00:01.096)       0:08:31.604 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Saturday 21 March 2026  00:53:10 +0000 (0:00:00.772)       0:08:32.377 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Saturday 21 March 2026  00:53:11 +0000 (0:00:00.822)       0:08:33.199 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Saturday 21 March 2026  00:53:12 +0000 (0:00:00.671)       0:08:33.871 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Saturday 21 March 2026  00:53:13 +0000 (0:00:01.190)       0:08:35.062 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Saturday 21 March 2026  00:53:14 +0000 (0:00:00.663)       0:08:35.725 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Saturday 21 March 2026  00:53:14 +0000 (0:00:00.633)       0:08:36.359 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Saturday 21 March 2026  00:53:16 +0000 (0:00:01.496)       0:08:37.855 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Saturday 21 March 2026  00:53:17 +0000 (0:00:01.193)       0:08:39.048 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Saturday 21 March 2026  00:53:18 +0000 (0:00:00.952)       0:08:40.000 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Saturday 21 March 2026  00:53:19 +0000 (0:00:00.548)       0:08:40.549 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Saturday 21 March 2026  00:53:19 +0000 (0:00:00.728)       0:08:41.277 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Saturday 21 March 2026  00:53:20 +0000 (0:00:00.519)       0:08:41.797 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Saturday 21 March 2026  00:53:21 +0000 (0:00:00.839)       0:08:42.637 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Saturday 21 March 2026  00:53:21 +0000 (0:00:00.525)       0:08:43.162 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Saturday 21 March 2026  00:53:22 +0000 (0:00:00.767)       0:08:43.930 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Saturday 21 March 2026  00:53:23 +0000 (0:00:00.619)       0:08:44.549 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Saturday 21 March 2026  00:53:23 +0000 (0:00:00.803)       0:08:45.352 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Create client.crash keyring] ********************************",
                                "Saturday 21 March 2026  00:53:24 +0000 (0:00:01.017)       0:08:46.370 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Get keys from monitors] *************************************",
                                "Saturday 21 March 2026  00:53:29 +0000 (0:00:04.055)       0:08:50.425 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Copy ceph key(s) if needed] *********************************",
                                "Saturday 21 March 2026  00:53:31 +0000 (0:00:02.105)       0:08:52.531 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Create /var/lib/ceph/crash/posted] **************************",
                                "Saturday 21 March 2026  00:53:33 +0000 (0:00:02.031)       0:08:54.562 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Include_tasks systemd.yml] **********************************",
                                "Saturday 21 March 2026  00:53:34 +0000 (0:00:01.488)       0:08:56.051 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-crash/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-crash : Generate systemd unit file for ceph-crash container] ********",
                                "Saturday 21 March 2026  00:53:35 +0000 (0:00:00.904)       0:08:56.955 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Start the ceph-crash service] *******************************",
                                "Saturday 21 March 2026  00:53:37 +0000 (0:00:01.767)       0:08:58.722 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Ceph crash handler] ****************************",
                                "Saturday 21 March 2026  00:53:41 +0000 (0:00:03.948)       0:09:02.671 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_crash.yml for testbed-node-4, testbed-node-3, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _crash_handler_called before restart] ******",
                                "Saturday 21 March 2026  00:53:42 +0000 (0:00:01.434)       0:09:04.106 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart the ceph-crash service] ****************",
                                "Saturday 21 March 2026  00:53:43 +0000 (0:00:00.694)       0:09:04.800 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _crash_handler_called after restart] *******",
                                "Saturday 21 March 2026  00:53:46 +0000 (0:00:03.137)       0:09:07.938 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-mds] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Saturday 21 March 2026  00:53:47 +0000 (0:00:00.885)       0:09:08.823 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Saturday 21 March 2026  00:53:48 +0000 (0:00:00.855)       0:09:09.678 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Saturday 21 March 2026  00:53:48 +0000 (0:00:00.572)       0:09:10.251 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Saturday 21 March 2026  00:53:49 +0000 (0:00:00.793)       0:09:11.045 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Saturday 21 March 2026  00:53:50 +0000 (0:00:00.819)       0:09:11.865 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Saturday 21 March 2026  00:53:51 +0000 (0:00:00.717)       0:09:12.582 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Saturday 21 March 2026  00:53:51 +0000 (0:00:00.780)       0:09:13.362 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Saturday 21 March 2026  00:53:52 +0000 (0:00:00.785)       0:09:14.148 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Saturday 21 March 2026  00:53:53 +0000 (0:00:00.342)       0:09:14.491 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Saturday 21 March 2026  00:53:53 +0000 (0:00:00.319)       0:09:14.810 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Saturday 21 March 2026  00:53:54 +0000 (0:00:00.680)       0:09:15.491 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Saturday 21 March 2026  00:53:55 +0000 (0:00:01.091)       0:09:16.583 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Saturday 21 March 2026  00:53:55 +0000 (0:00:00.307)       0:09:16.891 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Saturday 21 March 2026  00:53:55 +0000 (0:00:00.339)       0:09:17.230 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Saturday 21 March 2026  00:53:56 +0000 (0:00:00.434)       0:09:17.665 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Saturday 21 March 2026  00:53:57 +0000 (0:00:00.807)       0:09:18.472 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Saturday 21 March 2026  00:53:57 +0000 (0:00:00.385)       0:09:18.857 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Saturday 21 March 2026  00:53:57 +0000 (0:00:00.344)       0:09:19.202 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Saturday 21 March 2026  00:53:58 +0000 (0:00:00.373)       0:09:19.575 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Saturday 21 March 2026  00:53:58 +0000 (0:00:00.622)       0:09:20.198 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Saturday 21 March 2026  00:53:59 +0000 (0:00:00.344)       0:09:20.542 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Include create_mds_filesystems.yml] ***************************",
                                "Saturday 21 March 2026  00:53:59 +0000 (0:00:00.581)       0:09:21.123 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/create_mds_filesystems.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current default crush rule details] *********************",
                                "Saturday 21 March 2026  00:54:00 +0000 (0:00:00.629)       0:09:21.753 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current default crush rule name] ************************",
                                "Saturday 21 March 2026  00:54:02 +0000 (0:00:02.036)       0:09:23.790 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'rule_id': 0, 'rule_name': 'replicated_rule', 'type': 1, 'steps': [{'op': 'take', 'item': -1, 'item_name': 'default'}, {'op': 'chooseleaf_firstn', 'num': 0, 'type': 'host'}, {'op': 'emit'}]}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Create filesystem pools] **************************************",
                                "Saturday 21 March 2026  00:54:02 +0000 (0:00:00.224)       0:09:24.014 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'application': 'cephfs', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'cephfs_data', 'pg_num': 16, 'pgp_num': 16, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'application': 'cephfs', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'cephfs_metadata', 'pg_num': 16, 'pgp_num': 16, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "",
                                "TASK [ceph-mds : Create ceph filesystem] ***************************************",
                                "Saturday 21 March 2026  00:54:11 +0000 (0:00:09.263)       0:09:33.277 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Include common.yml] *******************************************",
                                "Saturday 21 March 2026  00:54:15 +0000 (0:00:03.277)       0:09:36.555 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-mds : Create bootstrap-mds and mds directories] *********************",
                                "Saturday 21 March 2026  00:54:15 +0000 (0:00:00.793)       0:09:37.349 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mds/ceph-testbed-node-3)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mds/ceph-testbed-node-4)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mds/ceph-testbed-node-5)\u001b[0m",
                                "",
                                "TASK [ceph-mds : Get keys from monitors] ***************************************",
                                "Saturday 21 March 2026  00:54:17 +0000 (0:00:01.274)       0:09:38.623 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Copy ceph key(s) if needed] ***********************************",
                                "Saturday 21 March 2026  00:54:19 +0000 (0:00:02.350)       0:09:40.974 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Create mds keyring] *******************************************",
                                "Saturday 21 March 2026  00:54:20 +0000 (0:00:01.046)       0:09:42.020 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Non_containerized.yml] ****************************************",
                                "Saturday 21 March 2026  00:54:22 +0000 (0:00:02.284)       0:09:44.304 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Containerized.yml] ********************************************",
                                "Saturday 21 March 2026  00:54:23 +0000 (0:00:00.257)       0:09:44.562 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/containerized.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-mds : Include_tasks systemd.yml] ************************************",
                                "Saturday 21 March 2026  00:54:23 +0000 (0:00:00.484)       0:09:45.047 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-mds : Generate systemd unit file] ***********************************",
                                "Saturday 21 March 2026  00:54:24 +0000 (0:00:00.634)       0:09:45.681 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Generate systemd ceph-mds target file] ************************",
                                "Saturday 21 March 2026  00:54:25 +0000 (0:00:01.287)       0:09:46.969 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Enable ceph-mds.target] ***************************************",
                                "Saturday 21 March 2026  00:54:26 +0000 (0:00:01.374)       0:09:48.343 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Systemd start mds container] **********************************",
                                "Saturday 21 March 2026  00:54:28 +0000 (0:00:01.971)       0:09:50.315 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Wait for mds socket to exist] *********************************",
                                "Saturday 21 March 2026  00:54:31 +0000 (0:00:02.452)       0:09:52.767 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Saturday 21 March 2026  00:54:32 +0000 (0:00:01.101)       0:09:53.868 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mdss handler] **********************************",
                                "Saturday 21 March 2026  00:54:33 +0000 (0:00:00.868)       0:09:54.736 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mdss.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mds_handler_called before restart] ********",
                                "Saturday 21 March 2026  00:54:33 +0000 (0:00:00.474)       0:09:55.211 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mds restart script] ***********************",
                                "Saturday 21 March 2026  00:54:34 +0000 (0:00:00.286)       0:09:55.497 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mds daemon(s)] ********************",
                                "Saturday 21 March 2026  00:54:35 +0000 (0:00:01.416)       0:09:56.914 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mds_handler_called after restart] *********",
                                "Saturday 21 March 2026  00:54:36 +0000 (0:00:00.676)       0:09:57.591 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-rgw] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Saturday 21 March 2026  00:54:36 +0000 (0:00:00.534)       0:09:58.125 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Saturday 21 March 2026  00:54:37 +0000 (0:00:00.701)       0:09:58.827 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Saturday 21 March 2026  00:54:37 +0000 (0:00:00.483)       0:09:59.311 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Saturday 21 March 2026  00:54:38 +0000 (0:00:00.420)       0:09:59.731 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Saturday 21 March 2026  00:54:39 +0000 (0:00:00.710)       0:10:00.441 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Saturday 21 March 2026  00:54:39 +0000 (0:00:00.734)       0:10:01.176 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Saturday 21 March 2026  00:54:40 +0000 (0:00:00.746)       0:10:01.922 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Saturday 21 March 2026  00:54:41 +0000 (0:00:00.601)       0:10:02.524 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Saturday 21 March 2026  00:54:41 +0000 (0:00:00.306)       0:10:02.830 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Saturday 21 March 2026  00:54:41 +0000 (0:00:00.292)       0:10:03.123 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Saturday 21 March 2026  00:54:42 +0000 (0:00:00.668)       0:10:03.791 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Saturday 21 March 2026  00:54:43 +0000 (0:00:00.978)       0:10:04.769 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Saturday 21 March 2026  00:54:43 +0000 (0:00:00.364)       0:10:05.133 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Saturday 21 March 2026  00:54:44 +0000 (0:00:00.337)       0:10:05.471 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Saturday 21 March 2026  00:54:44 +0000 (0:00:00.355)       0:10:05.826 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Saturday 21 March 2026  00:54:44 +0000 (0:00:00.573)       0:10:06.400 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Saturday 21 March 2026  00:54:45 +0000 (0:00:00.339)       0:10:06.740 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Saturday 21 March 2026  00:54:45 +0000 (0:00:00.360)       0:10:07.100 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Saturday 21 March 2026  00:54:46 +0000 (0:00:00.367)       0:10:07.468 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Saturday 21 March 2026  00:54:46 +0000 (0:00:00.634)       0:10:08.102 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Saturday 21 March 2026  00:54:47 +0000 (0:00:00.370)       0:10:08.473 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include common.yml] *******************************************",
                                "Saturday 21 March 2026  00:54:47 +0000 (0:00:00.527)       0:10:09.000 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Get keys from monitors] ***************************************",
                                "Saturday 21 March 2026  00:54:48 +0000 (0:00:00.921)       0:10:09.921 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************",
                                "Saturday 21 March 2026  00:54:50 +0000 (0:00:02.124)       0:10:12.046 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Copy SSL certificate & key data to certificate path] **********",
                                "Saturday 21 March 2026  00:54:51 +0000 (0:00:01.082)       0:10:13.129 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_tasks pre_requisite.yml] ******************************",
                                "Saturday 21 March 2026  00:54:52 +0000 (0:00:00.333)       0:10:13.463 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/pre_requisite.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create rados gateway directories] *****************************",
                                "Saturday 21 March 2026  00:54:52 +0000 (0:00:00.801)       0:10:14.264 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create rgw keyrings] ******************************************",
                                "Saturday 21 March 2026  00:54:53 +0000 (0:00:00.787)       0:10:15.051 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Get keys from monitors] ***************************************",
                                "Saturday 21 March 2026  00:54:57 +0000 (0:00:04.085)       0:10:19.136 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************",
                                "Saturday 21 March 2026  00:55:00 +0000 (0:00:02.352)       0:10:21.488 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Rgw pool creation tasks] **************************************",
                                "Saturday 21 March 2026  00:55:01 +0000 (0:00:01.111)       0:10:22.600 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/rgw_create_pools.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create ec profile] ********************************************",
                                "Saturday 21 March 2026  00:55:01 +0000 (0:00:00.232)       0:10:22.832 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Set crush rule] ***********************************************",
                                "Saturday 21 March 2026  00:55:01 +0000 (0:00:00.566)       0:10:23.398 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create rgw pools] *********************************************",
                                "Saturday 21 March 2026  00:55:02 +0000 (0:00:00.616)       0:10:24.015 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_tasks openstack-keystone.yml] *************************",
                                "Saturday 21 March 2026  00:55:32 +0000 (0:00:29.874)       0:10:53.890 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_tasks start_radosgw.yml] ******************************",
                                "Saturday 21 March 2026  00:55:32 +0000 (0:00:00.326)       0:10:54.216 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include start_docker_rgw.yml] *********************************",
                                "Saturday 21 March 2026  00:55:33 +0000 (0:00:00.637)       0:10:54.854 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/start_docker_rgw.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_task systemd.yml] *************************************",
                                "Saturday 21 March 2026  00:55:33 +0000 (0:00:00.553)       0:10:55.407 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Generate systemd unit file] ***********************************",
                                "Saturday 21 March 2026  00:55:34 +0000 (0:00:00.774)       0:10:56.182 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Generate systemd ceph-radosgw target file] ********************",
                                "Saturday 21 March 2026  00:55:35 +0000 (0:00:01.148)       0:10:57.330 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Enable ceph-radosgw.target] ***********************************",
                                "Saturday 21 March 2026  00:55:36 +0000 (0:00:01.078)       0:10:58.409 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Systemd start rgw container] **********************************",
                                "Saturday 21 March 2026  00:55:38 +0000 (0:00:01.663)       0:11:00.072 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Saturday 21 March 2026  00:55:41 +0000 (0:00:02.415)       0:11:02.488 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rgws handler] **********************************",
                                "Saturday 21 March 2026  00:55:41 +0000 (0:00:00.305)       0:11:02.794 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_rgws.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _rgw_handler_called before restart] ********",
                                "Saturday 21 March 2026  00:55:42 +0000 (0:00:00.805)       0:11:03.599 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy rgw restart script] ***********************",
                                "Saturday 21 March 2026  00:55:42 +0000 (0:00:00.315)       0:11:03.915 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph rgw daemon(s)] ********************",
                                "Saturday 21 March 2026  00:55:42 +0000 (0:00:00.308)       0:11:04.223 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _rgw_handler_called after restart] *********",
                                "Saturday 21 March 2026  00:55:43 +0000 (0:00:01.116)       0:11:05.339 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=134 \u001b[0m \u001b[0;33mchanged=35  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=125 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=127 \u001b[0m \u001b[0;33mchanged=31  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=120 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=134 \u001b[0m \u001b[0;33mchanged=33  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=119 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=193 \u001b[0m \u001b[0;33mchanged=45  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=162 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=40  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=123 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=41  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=121 \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:55:44 +0000 (0:00:00.245)       0:11:05.585 ******** ",
                                "=============================================================================== ",
                                "ceph-container-common : Pulling Ceph container image ------------------- 58.61s",
                                "ceph-osd : Use ceph-volume to create osds ------------------------------ 44.17s",
                                "ceph-mgr : Wait for all mgr to be up ----------------------------------- 36.71s",
                                "ceph-rgw : Create rgw pools -------------------------------------------- 29.88s",
                                "ceph-mon : Waiting for the monitor(s) to form the quorum... ------------ 21.52s",
                                "ceph-mon : Set cluster configs ----------------------------------------- 14.96s",
                                "ceph-osd : Wait for all osd to be up ----------------------------------- 12.86s",
                                "ceph-mgr : Create ceph mgr keyring(s) on a mon node -------------------- 10.78s",
                                "ceph-mds : Create filesystem pools -------------------------------------- 9.26s",
                                "ceph-mon : Fetch ceph initial keys -------------------------------------- 8.91s",
                                "ceph-config : Create ceph initial directories --------------------------- 6.69s",
                                "ceph-mgr : Disable ceph mgr enabled modules ----------------------------- 6.39s",
                                "ceph-mgr : Add modules to ceph-mgr -------------------------------------- 4.81s",
                                "ceph-rgw : Create rgw keyrings ------------------------------------------ 4.09s",
                                "ceph-crash : Create client.crash keyring -------------------------------- 4.06s",
                                "ceph-mon : Copy admin keyring over to mons ------------------------------ 4.01s",
                                "ceph-osd : Apply operating system tuning -------------------------------- 3.96s",
                                "ceph-crash : Start the ceph-crash service ------------------------------- 3.95s",
                                "ceph-osd : Systemd start osd -------------------------------------------- 3.86s",
                                "ceph-facts : Set_fact _monitor_addresses - ipv4 ------------------------- 3.59s",
                                "\u001b[32m2026-03-21 00:55:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:55:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b439d680-2298-49e1-9481-b754420390bd is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Set kolla_action_mariadb] ************************************************",
                                "",
                                "TASK [Inform the user about the following task] ********************************",
                                "Saturday 21 March 2026  00:53:10 +0000 (0:00:00.092)       0:00:00.092 ******** ",
                                "\u001b[0;32mok: [localhost] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"The task 'Check MariaDB service' fails if the MariaDB service has not yet been deployed. This is fine.\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Check MariaDB service] ***************************************************",
                                "Saturday 21 March 2026  00:53:11 +0000 (0:00:00.035)       0:00:00.128 ******** ",
                                "\u001b[0;31mfatal: [localhost]: FAILED! => {\"changed\": false, \"elapsed\": 2, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.9:3306\"}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "",
                                "TASK [Set kolla_action_mariadb = upgrade if MariaDB is already running] ********",
                                "Saturday 21 March 2026  00:53:13 +0000 (0:00:02.814)       0:00:02.942 ******** ",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "TASK [Set kolla_action_mariadb = kolla_action_ng] ******************************",
                                "Saturday 21 March 2026  00:53:13 +0000 (0:00:00.052)       0:00:02.995 ******** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  00:53:14 +0000 (0:00:00.292)       0:00:03.287 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  00:53:14 +0000 (0:00:00.305)       0:00:03.593 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_mariadb_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_mariadb_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_mariadb_True)\u001b[0m",
                                "",
                                "PLAY [Apply role mariadb] ******************************************************",
                                "",
                                "TASK [mariadb : Group MariaDB hosts based on shards] ***************************",
                                "Saturday 21 March 2026  00:53:15 +0000 (0:00:00.566)       0:00:04.159 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Saturday 21 March 2026  00:53:15 +0000 (0:00:00.501)       0:00:04.661 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [mariadb : Ensuring config directories exist] *****************************",
                                "Saturday 21 March 2026  00:53:16 +0000 (0:00:00.965)       0:00:05.626 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "",
                                "TASK [mariadb : Ensuring database backup config directory exists] **************",
                                "Saturday 21 March 2026  00:53:19 +0000 (0:00:03.311)       0:00:08.938 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Copying over my.cnf for mariabackup] ***************************",
                                "Saturday 21 March 2026  00:53:20 +0000 (0:00:00.756)       0:00:09.694 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Copying over config.json files for services] *******************",
                                "Saturday 21 March 2026  00:53:21 +0000 (0:00:01.415)       0:00:11.110 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "",
                                "TASK [mariadb : Copying over config.json files for mariabackup] ****************",
                                "Saturday 21 March 2026  00:53:26 +0000 (0:00:04.193)       0:00:15.303 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Copying over galera.cnf] ***************************************",
                                "Saturday 21 March 2026  00:53:27 +0000 (0:00:01.093)       0:00:16.397 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Saturday 21 March 2026  00:53:30 +0000 (0:00:03.376)       0:00:19.774 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over extra CA certificates] ********",
                                "Saturday 21 March 2026  00:53:31 +0000 (0:00:00.488)       0:00:20.263 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] ***",
                                "Saturday 21 March 2026  00:53:34 +0000 (0:00:03.419)       0:00:23.682 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over backend internal TLS key] *****",
                                "Saturday 21 March 2026  00:53:36 +0000 (0:00:02.288)       0:00:25.970 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Check mariadb containers] **************************************",
                                "Saturday 21 March 2026  00:53:39 +0000 (0:00:02.711)       0:00:28.682 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "",
                                "TASK [mariadb : Create MariaDB volume] *****************************************",
                                "Saturday 21 March 2026  00:53:43 +0000 (0:00:03.503)       0:00:32.185 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Divide hosts by their MariaDB volume availability] *************",
                                "Saturday 21 March 2026  00:53:43 +0000 (0:00:00.811)       0:00:32.997 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Establish whether the cluster has already existed] *************",
                                "Saturday 21 March 2026  00:53:44 +0000 (0:00:00.349)       0:00:33.347 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Check MariaDB service port liveness] ***************************",
                                "Saturday 21 March 2026  00:53:44 +0000 (0:00:00.328)       0:00:33.675 ******** ",
                                "\u001b[0;31mfatal: [testbed-node-0]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.10:3306\"}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "\u001b[0;31mfatal: [testbed-node-1]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.11:3306\"}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "\u001b[0;31mfatal: [testbed-node-2]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.12:3306\"}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "",
                                "TASK [mariadb : Divide hosts by their MariaDB service port liveness] ***********",
                                "Saturday 21 March 2026  00:53:55 +0000 (0:00:11.221)       0:00:44.896 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Fail on existing but stopped cluster] **************************",
                                "Saturday 21 March 2026  00:53:56 +0000 (0:00:00.492)       0:00:45.389 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Check MariaDB service WSREP sync status] ***********************",
                                "Saturday 21 March 2026  00:53:56 +0000 (0:00:00.440)       0:00:45.830 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Extract MariaDB service WSREP sync status] *********************",
                                "Saturday 21 March 2026  00:53:57 +0000 (0:00:00.515)       0:00:46.345 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Divide hosts by their MariaDB service WSREP sync status] *******",
                                "Saturday 21 March 2026  00:53:57 +0000 (0:00:00.694)       0:00:47.039 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Fail when MariaDB services are not synced across the whole cluster] ***",
                                "Saturday 21 March 2026  00:53:58 +0000 (0:00:00.486)       0:00:47.525 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Saturday 21 March 2026  00:53:58 +0000 (0:00:00.457)       0:00:47.983 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/bootstrap_cluster.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [mariadb : Running MariaDB bootstrap container] ***************************",
                                "Saturday 21 March 2026  00:53:59 +0000 (0:00:00.398)       0:00:48.382 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Store bootstrap host name into facts] **************************",
                                "Saturday 21 March 2026  00:54:09 +0000 (0:00:10.698)       0:00:59.080 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Saturday 21 March 2026  00:54:10 +0000 (0:00:00.274)       0:00:59.355 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Starting first MariaDB container] *******************",
                                "Saturday 21 March 2026  00:54:11 +0000 (0:00:00.859)       0:01:00.214 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Wait for first MariaDB service port liveness] *******",
                                "Saturday 21 March 2026  00:54:18 +0000 (0:00:07.587)       0:01:07.802 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Wait for first MariaDB service to sync WSREP] *******",
                                "Saturday 21 March 2026  00:54:20 +0000 (0:00:01.528)       0:01:09.331 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Ensure MariaDB is running normally on bootstrap host] ***",
                                "Saturday 21 March 2026  00:54:22 +0000 (0:00:01.901)       0:01:11.232 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Restart MariaDB on existing cluster members] ********",
                                "Saturday 21 March 2026  00:54:22 +0000 (0:00:00.220)       0:01:11.453 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Start MariaDB on new nodes] *************************",
                                "Saturday 21 March 2026  00:54:22 +0000 (0:00:00.297)       0:01:11.751 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_restart\u001b[0m",
                                "",
                                "PLAY [Restart mariadb services] ************************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Start mariadb services] **************************************************",
                                "",
                                "TASK [mariadb : Restart MariaDB container] *************************************",
                                "Saturday 21 March 2026  00:54:22 +0000 (0:00:00.276)       0:01:12.027 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service port liveness] ************************",
                                "Saturday 21 March 2026  00:54:44 +0000 (0:00:21.805)       0:01:33.833 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to sync WSREP] ************************",
                                "Saturday 21 March 2026  00:54:54 +0000 (0:00:09.593)       0:01:43.426 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Start mariadb services] **************************************************",
                                "",
                                "TASK [mariadb : Restart MariaDB container] *************************************",
                                "Saturday 21 March 2026  00:54:56 +0000 (0:00:02.447)       0:01:45.874 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service port liveness] ************************",
                                "Saturday 21 March 2026  00:55:13 +0000 (0:00:17.000)       0:02:02.875 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to sync WSREP] ************************",
                                "Saturday 21 March 2026  00:55:28 +0000 (0:00:14.938)       0:02:17.813 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Restart bootstrap mariadb service] ***************************************",
                                "",
                                "TASK [mariadb : Restart MariaDB container] *************************************",
                                "Saturday 21 March 2026  00:55:31 +0000 (0:00:02.651)       0:02:20.465 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service port liveness] ************************",
                                "Saturday 21 March 2026  00:55:42 +0000 (0:00:11.109)       0:02:31.574 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to sync WSREP] ************************",
                                "Saturday 21 March 2026  00:55:45 +0000 (0:00:03.486)       0:02:35.061 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Apply mariadb post-configuration] ****************************************",
                                "",
                                "TASK [Include mariadb post-deploy.yml] *****************************************",
                                "Saturday 21 March 2026  00:55:48 +0000 (0:00:02.445)       0:02:37.506 ******** ",
                                "\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [mariadb : Creating shard root mysql user] ********************************",
                                "Saturday 21 March 2026  00:55:49 +0000 (0:00:00.662)       0:02:38.168 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Creating mysql monitor user] ***********************************",
                                "Saturday 21 March 2026  00:55:51 +0000 (0:00:02.281)       0:02:40.450 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Creating database backup user and setting permissions] *********",
                                "Saturday 21 March 2026  00:55:53 +0000 (0:00:01.902)       0:02:42.352 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Granting permissions on Mariabackup database to backup user] ***",
                                "Saturday 21 March 2026  00:55:55 +0000 (0:00:01.849)       0:02:44.202 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to be ready through VIP] **************",
                                "Saturday 21 March 2026  00:55:57 +0000 (0:00:02.324)       0:02:46.526 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Include mariadb post-upgrade.yml] ****************************************",
                                "Saturday 21 March 2026  00:56:00 +0000 (0:00:02.833)       0:02:49.360 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mlocalhost\u001b[0m                  : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=18  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=18  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:56:00 +0000 (0:00:00.244)       0:02:49.604 ******** ",
                                "=============================================================================== ",
                                "mariadb : Restart MariaDB container ------------------------------------ 38.81s",
                                "mariadb : Wait for MariaDB service port liveness ----------------------- 24.53s",
                                "mariadb : Check MariaDB service port liveness -------------------------- 11.22s",
                                "mariadb : Restart MariaDB container ------------------------------------ 11.11s",
                                "mariadb : Running MariaDB bootstrap container -------------------------- 10.70s",
                                "mariadb : Starting first MariaDB container ------------------------------ 7.59s",
                                "mariadb : Wait for MariaDB service to sync WSREP ------------------------ 5.10s",
                                "mariadb : Copying over config.json files for services ------------------- 4.19s",
                                "mariadb : Check mariadb containers -------------------------------------- 3.50s",
                                "mariadb : Wait for MariaDB service port liveness ------------------------ 3.49s",
                                "service-cert-copy : mariadb | Copying over extra CA certificates -------- 3.42s",
                                "mariadb : Copying over galera.cnf --------------------------------------- 3.38s",
                                "mariadb : Ensuring config directories exist ----------------------------- 3.31s",
                                "mariadb : Wait for MariaDB service to be ready through VIP -------------- 2.83s",
                                "Check MariaDB service --------------------------------------------------- 2.81s",
                                "service-cert-copy : mariadb | Copying over backend internal TLS key ----- 2.71s",
                                "mariadb : Wait for MariaDB service to sync WSREP ------------------------ 2.45s",
                                "mariadb : Granting permissions on Mariabackup database to backup user --- 2.32s",
                                "service-cert-copy : mariadb | Copying over backend internal TLS certificate --- 2.29s",
                                "mariadb : Creating shard root mysql user -------------------------------- 2.28s",
                                "\u001b[32m2026-03-21 00:56:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:56:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 478e369f-207e-4a55-a9a3-cb8732dc5d74 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  00:56:03 +0000 (0:00:00.282)       0:00:00.282 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  00:56:04 +0000 (0:00:00.243)       0:00:00.525 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_horizon_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_horizon_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_horizon_True)\u001b[0m",
                                "",
                                "PLAY [Apply role horizon] ******************************************************",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Saturday 21 March 2026  00:56:04 +0000 (0:00:00.279)       0:00:00.805 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [horizon : Ensuring config directories exist] *****************************",
                                "Saturday 21 March 2026  00:56:04 +0000 (0:00:00.441)       0:00:01.247 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "",
                                "TASK [horizon : Set empty custom policy] ***************************************",
                                "Saturday 21 March 2026  00:56:06 +0000 (0:00:01.774)       0:00:03.021 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Saturday 21 March 2026  00:56:06 +0000 (0:00:00.260)       0:00:03.282 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'trove', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'trove', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'trove', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'ceilometer', 'enabled': 'yes'})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'cinder', 'enabled': 'yes'})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'designate', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'glance', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'keystone', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'magnum', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'manila', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'neutron', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'nova', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'octavia', 'enabled': True})\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Saturday 21 March 2026  00:56:07 +0000 (0:00:00.630)       0:00:03.912 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Saturday 21 March 2026  00:56:07 +0000 (0:00:00.370)       0:00:04.283 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Saturday 21 March 2026  00:56:07 +0000 (0:00:00.140)       0:00:04.423 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Saturday 21 March 2026  00:56:08 +0000 (0:00:00.238)       0:00:04.662 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Saturday 21 March 2026  00:56:08 +0000 (0:00:00.273)       0:00:04.936 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Saturday 21 March 2026  00:56:08 +0000 (0:00:00.115)       0:00:05.052 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Saturday 21 March 2026  00:56:08 +0000 (0:00:00.365)       0:00:05.417 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Saturday 21 March 2026  00:56:09 +0000 (0:00:00.387)       0:00:05.804 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Saturday 21 March 2026  00:56:09 +0000 (0:00:00.179)       0:00:05.984 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Saturday 21 March 2026  00:56:09 +0000 (0:00:00.329)       0:00:06.313 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Saturday 21 March 2026  00:56:10 +0000 (0:00:00.260)       0:00:06.574 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Saturday 21 March 2026  00:56:10 +0000 (0:00:00.118)       0:00:06.693 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Saturday 21 March 2026  00:56:10 +0000 (0:00:00.424)       0:00:07.118 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Saturday 21 March 2026  00:56:10 +0000 (0:00:00.283)       0:00:07.401 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Saturday 21 March 2026  00:56:11 +0000 (0:00:00.101)       0:00:07.503 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Saturday 21 March 2026  00:56:11 +0000 (0:00:00.251)       0:00:07.755 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Saturday 21 March 2026  00:56:11 +0000 (0:00:00.395)       0:00:08.151 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Saturday 21 March 2026  00:56:11 +0000 (0:00:00.101)       0:00:08.253 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Saturday 21 March 2026  00:56:12 +0000 (0:00:00.256)       0:00:08.509 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Saturday 21 March 2026  00:56:12 +0000 (0:00:00.366)       0:00:08.876 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Saturday 21 March 2026  00:56:12 +0000 (0:00:00.089)       0:00:08.966 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Saturday 21 March 2026  00:56:12 +0000 (0:00:00.250)       0:00:09.217 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Saturday 21 March 2026  00:56:13 +0000 (0:00:00.416)       0:00:09.633 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Saturday 21 March 2026  00:56:13 +0000 (0:00:00.102)       0:00:09.736 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Saturday 21 March 2026  00:56:13 +0000 (0:00:00.253)       0:00:09.989 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Saturday 21 March 2026  00:56:14 +0000 (0:00:00.484)       0:00:10.474 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Saturday 21 March 2026  00:56:14 +0000 (0:00:00.139)       0:00:10.614 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Saturday 21 March 2026  00:56:14 +0000 (0:00:00.269)       0:00:10.883 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Saturday 21 March 2026  00:56:14 +0000 (0:00:00.480)       0:00:11.364 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Saturday 21 March 2026  00:56:15 +0000 (0:00:00.126)       0:00:11.491 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Copying over config.json files for services] *******************",
                                "Saturday 21 March 2026  00:56:15 +0000 (0:00:00.289)       0:00:11.781 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Copying over horizon.conf] *************************************",
                                "Saturday 21 March 2026  00:56:17 +0000 (0:00:01.903)       0:00:13.684 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m",
                                "",
                                "TASK [horizon : Copying over kolla-settings.py] ********************************",
                                "Saturday 21 March 2026  00:56:19 +0000 (0:00:02.765)       0:00:16.450 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m",
                                "",
                                "TASK [horizon : Copying over custom-settings.py] *******************************",
                                "Saturday 21 March 2026  00:56:21 +0000 (0:00:01.935)       0:00:18.386 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m",
                                "",
                                "TASK [horizon : Copying over existing policy file] *****************************",
                                "Saturday 21 March 2026  00:56:23 +0000 (0:00:01.749)       0:00:20.135 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Copying over custom themes] ************************************",
                                "Saturday 21 March 2026  00:56:23 +0000 (0:00:00.282)       0:00:20.417 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Saturday 21 March 2026  00:56:24 +0000 (0:00:00.277)       0:00:20.695 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : horizon | Copying over extra CA certificates] ********",
                                "Saturday 21 March 2026  00:56:25 +0000 (0:00:00.844)       0:00:21.540 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : horizon | Copying over backend internal TLS certificate] ***",
                                "Saturday 21 March 2026  00:56:26 +0000 (0:00:01.593)       0:00:23.133 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : horizon | Copying over backend internal TLS key] *****",
                                "Saturday 21 March 2026  00:56:27 +0000 (0:00:00.927)       0:00:24.060 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Deploy horizon container] **************************************",
                                "Saturday 21 March 2026  00:56:28 +0000 (0:00:01.000)       0:00:25.060 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Saturday 21 March 2026  00:56:29 +0000 (0:00:01.381)       0:00:26.442 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Saturday 21 March 2026  00:56:30 +0000 (0:00:00.307)       0:00:26.749 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/bootstrap.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [horizon : Creating Horizon database] *************************************",
                                "Saturday 21 March 2026  00:56:31 +0000 (0:00:00.724)       0:00:27.474 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Creating Horizon database user and setting permissions] ********",
                                "Saturday 21 March 2026  00:56:33 +0000 (0:00:02.693)       0:00:30.168 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Running Horizon bootstrap container] ***************************",
                                "Saturday 21 March 2026  00:56:35 +0000 (0:00:02.052)       0:00:32.220 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Flush handlers] ************************************************",
                                "Saturday 21 March 2026  00:56:50 +0000 (0:00:15.090)       0:00:47.310 ******** ",
                                "",
                                "TASK [horizon : Flush handlers] ************************************************",
                                "Saturday 21 March 2026  00:56:50 +0000 (0:00:00.065)       0:00:47.376 ******** ",
                                "",
                                "TASK [horizon : Flush handlers] ************************************************",
                                "Saturday 21 March 2026  00:56:50 +0000 (0:00:00.081)       0:00:47.457 ******** ",
                                "",
                                "RUNNING HANDLER [horizon : Restart horizon container] **************************",
                                "Saturday 21 March 2026  00:56:51 +0000 (0:00:00.088)       0:00:47.546 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=37  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=15  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=15  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:57:36 +0000 (0:00:45.880)       0:01:33.427 ******** ",
                                "=============================================================================== ",
                                "horizon : Restart horizon container ------------------------------------ 45.88s",
                                "horizon : Running Horizon bootstrap container -------------------------- 15.09s",
                                "horizon : Copying over horizon.conf ------------------------------------- 2.77s",
                                "horizon : Creating Horizon database ------------------------------------- 2.69s",
                                "horizon : Creating Horizon database user and setting permissions -------- 2.05s",
                                "horizon : Copying over kolla-settings.py -------------------------------- 1.94s",
                                "horizon : Copying over config.json files for services ------------------- 1.90s",
                                "horizon : Ensuring config directories exist ----------------------------- 1.77s",
                                "horizon : Copying over custom-settings.py ------------------------------- 1.75s",
                                "service-cert-copy : horizon | Copying over extra CA certificates -------- 1.59s",
                                "horizon : Deploy horizon container -------------------------------------- 1.38s",
                                "service-cert-copy : horizon | Copying over backend internal TLS key ----- 1.00s",
                                "service-cert-copy : horizon | Copying over backend internal TLS certificate --- 0.93s",
                                "horizon : include_tasks ------------------------------------------------- 0.84s",
                                "horizon : include_tasks ------------------------------------------------- 0.72s",
                                "horizon : include_tasks ------------------------------------------------- 0.63s",
                                "horizon : Update policy file name --------------------------------------- 0.48s",
                                "horizon : Update policy file name --------------------------------------- 0.48s",
                                "horizon : include_tasks ------------------------------------------------- 0.44s",
                                "horizon : Update custom policy file name -------------------------------- 0.42s",
                                "\u001b[32m2026-03-21 00:57:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d05e1ff-681e-41de-a7e8-fdaf6fa92820 is in state SUCCESS\u001b[0m",
                                "",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Create ceph pools] *******************************************************",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Saturday 21 March 2026  00:55:49 +0000 (0:00:00.551)       0:00:00.551 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Saturday 21 March 2026  00:55:49 +0000 (0:00:00.474)       0:00:01.025 ******** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Saturday 21 March 2026  00:55:50 +0000 (0:00:00.844)       0:00:01.869 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Saturday 21 March 2026  00:55:50 +0000 (0:00:00.251)       0:00:02.121 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Saturday 21 March 2026  00:55:51 +0000 (0:00:00.681)       0:00:02.802 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Saturday 21 March 2026  00:55:51 +0000 (0:00:00.291)       0:00:03.094 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Saturday 21 March 2026  00:55:52 +0000 (0:00:00.260)       0:00:03.355 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Saturday 21 March 2026  00:55:52 +0000 (0:00:00.252)       0:00:03.608 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Saturday 21 March 2026  00:55:52 +0000 (0:00:00.391)       0:00:03.999 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Saturday 21 March 2026  00:55:53 +0000 (0:00:00.286)       0:00:04.286 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Saturday 21 March 2026  00:55:53 +0000 (0:00:00.635)       0:00:04.922 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Saturday 21 March 2026  00:55:54 +0000 (0:00:00.407)       0:00:05.329 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Saturday 21 March 2026  00:55:56 +0000 (0:00:02.786)       0:00:08.115 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Saturday 21 March 2026  00:55:57 +0000 (0:00:00.419)       0:00:08.535 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Saturday 21 March 2026  00:55:58 +0000 (0:00:00.838)       0:00:09.374 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Saturday 21 March 2026  00:55:58 +0000 (0:00:00.186)       0:00:09.560 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': '22929cdeac3a', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-21 00:55:54.962395', 'end': '2026-03-21 00:55:54.995637', 'delta': '0:00:00.033242', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['22929cdeac3a'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': 'a96b2c5299e5', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-21 00:55:55.943631', 'end': '2026-03-21 00:55:55.973282', 'delta': '0:00:00.029651', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['a96b2c5299e5'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': '57e0bd4375ae', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-21 00:55:56.696837', 'end': '2026-03-21 00:55:56.725035', 'delta': '0:00:00.028198', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['57e0bd4375ae'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Saturday 21 March 2026  00:55:58 +0000 (0:00:00.376)       0:00:09.937 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Saturday 21 March 2026  00:55:59 +0000 (0:00:00.554)       0:00:10.491 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Saturday 21 March 2026  00:56:00 +0000 (0:00:01.599)       0:00:12.091 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Saturday 21 March 2026  00:56:01 +0000 (0:00:00.274)       0:00:12.365 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Saturday 21 March 2026  00:56:01 +0000 (0:00:00.433)       0:00:12.798 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Saturday 21 March 2026  00:56:02 +0000 (0:00:00.477)       0:00:13.276 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Saturday 21 March 2026  00:56:02 +0000 (0:00:00.121)       0:00:13.398 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Saturday 21 March 2026  00:56:02 +0000 (0:00:00.205)       0:00:13.604 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Saturday 21 March 2026  00:56:02 +0000 (0:00:00.238)       0:00:13.842 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Saturday 21 March 2026  00:56:02 +0000 (0:00:00.285)       0:00:14.127 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Saturday 21 March 2026  00:56:03 +0000 (0:00:00.403)       0:00:14.530 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Saturday 21 March 2026  00:56:03 +0000 (0:00:00.273)       0:00:14.804 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Saturday 21 March 2026  00:56:03 +0000 (0:00:00.269)       0:00:15.073 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Saturday 21 March 2026  00:56:04 +0000 (0:00:00.263)       0:00:15.337 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Saturday 21 March 2026  00:56:04 +0000 (0:00:00.394)       0:00:15.731 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--64dd100a--6bcc--5763--b557--0732c79bfd55-osd--block--64dd100a--6bcc--5763--b557--0732c79bfd55', 'dm-uuid-LVM-MWn87OlXxZ0hjMuGG89puNwiFbMviFcm1xAms0qPz8LYU8Ut0AGzeF9WoI61D5NM'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--5368e13a--d21f--5912--91bd--da537815d9e8-osd--block--5368e13a--d21f--5912--91bd--da537815d9e8', 'dm-uuid-LVM-xgLPNxLNUSwUnkWnyYv6gd71t6QIH1sH9rPUXF3date4d4aBKBICcuqoW7mNU6TI'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--ed04aa0b--4f5a--523b--a5d2--05072a1b260d-osd--block--ed04aa0b--4f5a--523b--a5d2--05072a1b260d', 'dm-uuid-LVM-viMZkgfP3jZUgTAo0bSRpd40fLtFkJJmO8XW5mmi1aRChVrq8X3DDkwGl8XRb5at'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part1', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part14', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part15', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part16', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--45b09425--beda--50b4--a83b--1fc50b5adca4-osd--block--45b09425--beda--50b4--a83b--1fc50b5adca4', 'dm-uuid-LVM-XzfTGfmLHoftdDBti3ot8l7sbRObO8yALj7bfZRysK7QQ9WjzIgwF2E0nO1Pc63W'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'holders': ['ceph--64dd100a--6bcc--5763--b557--0732c79bfd55-osd--block--64dd100a--6bcc--5763--b557--0732c79bfd55'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-HiPrG3-9qzo-WJZ0-1Smx-Bi1t-ReBz-6cdxta', 'scsi-0QEMU_QEMU_HARDDISK_f08311eb-e2ad-4eb7-8da9-ac6d899a2409', 'scsi-SQEMU_QEMU_HARDDISK_f08311eb-e2ad-4eb7-8da9-ac6d899a2409'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'holders': ['ceph--5368e13a--d21f--5912--91bd--da537815d9e8-osd--block--5368e13a--d21f--5912--91bd--da537815d9e8'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-m0F2nF-j9UB-mZfT-bRv8-xW1c-EVch-ZfDvHt', 'scsi-0QEMU_QEMU_HARDDISK_e03b171b-b829-41a0-a0b6-cc46d85f013d', 'scsi-SQEMU_QEMU_HARDDISK_e03b171b-b829-41a0-a0b6-cc46d85f013d'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_9fa3c153-0e2a-45ef-8e48-e0095eb937ad', 'scsi-SQEMU_QEMU_HARDDISK_9fa3c153-0e2a-45ef-8e48-e0095eb937ad'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-21-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part1', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part14', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part15', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part16', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'holders': ['ceph--ed04aa0b--4f5a--523b--a5d2--05072a1b260d-osd--block--ed04aa0b--4f5a--523b--a5d2--05072a1b260d'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-nra7ru-O0w6-dM26-lz5o-2arZ-rXsE-lANagX', 'scsi-0QEMU_QEMU_HARDDISK_4e74a544-8b3a-4a05-98a1-b7cc13f13c18', 'scsi-SQEMU_QEMU_HARDDISK_4e74a544-8b3a-4a05-98a1-b7cc13f13c18'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'holders': ['ceph--45b09425--beda--50b4--a83b--1fc50b5adca4-osd--block--45b09425--beda--50b4--a83b--1fc50b5adca4'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-XQTe4H-QTjK-DLSs-Qjj6-RKjC-da4T-xXChra', 'scsi-0QEMU_QEMU_HARDDISK_666e73fe-87f5-4f65-9ff2-95a038c8b18d', 'scsi-SQEMU_QEMU_HARDDISK_666e73fe-87f5-4f65-9ff2-95a038c8b18d'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ae4aabb-0174-4999-a1d0-4bf78cbb2965', 'scsi-SQEMU_QEMU_HARDDISK_8ae4aabb-0174-4999-a1d0-4bf78cbb2965'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-14-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--ede169e1--d974--55ed--a938--c303cb89cb5b-osd--block--ede169e1--d974--55ed--a938--c303cb89cb5b', 'dm-uuid-LVM-bEBXWAqtQdfT6xKYbL0dF10We9CLdKgKh0pf0vpR2lQnL405kp4ndXs76m3o0J5Y'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a-osd--block--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a', 'dm-uuid-LVM-G1LwYmzPCvwLZRWX1y5oYSHlzZdIxlOhfbzzvktDUwfRIgHaiaw6r0MV8HVh1TLk'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part1', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part14', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part15', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part16', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'holders': ['ceph--ede169e1--d974--55ed--a938--c303cb89cb5b-osd--block--ede169e1--d974--55ed--a938--c303cb89cb5b'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-HVhEyD-n8kM-PvHm-CUDD-tuRZ-igjC-Nzour2', 'scsi-0QEMU_QEMU_HARDDISK_f76ae8f0-65cc-4a62-b19c-23f8a40fbde6', 'scsi-SQEMU_QEMU_HARDDISK_f76ae8f0-65cc-4a62-b19c-23f8a40fbde6'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'holders': ['ceph--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a-osd--block--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-YifJdB-JYWE-HhRV-mF1g-0k83-hzWG-ctJZWy', 'scsi-0QEMU_QEMU_HARDDISK_d7bb81ba-6491-4b5a-bb82-f0b49314eac0', 'scsi-SQEMU_QEMU_HARDDISK_d7bb81ba-6491-4b5a-bb82-f0b49314eac0'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_cdffe5f1-7f0d-460e-a1db-d6846d226144', 'scsi-SQEMU_QEMU_HARDDISK_cdffe5f1-7f0d-460e-a1db-d6846d226144'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-28-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Saturday 21 March 2026  00:56:05 +0000 (0:00:00.551)       0:00:16.282 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--64dd100a--6bcc--5763--b557--0732c79bfd55-osd--block--64dd100a--6bcc--5763--b557--0732c79bfd55', 'dm-uuid-LVM-MWn87OlXxZ0hjMuGG89puNwiFbMviFcm1xAms0qPz8LYU8Ut0AGzeF9WoI61D5NM'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--5368e13a--d21f--5912--91bd--da537815d9e8-osd--block--5368e13a--d21f--5912--91bd--da537815d9e8', 'dm-uuid-LVM-xgLPNxLNUSwUnkWnyYv6gd71t6QIH1sH9rPUXF3date4d4aBKBICcuqoW7mNU6TI'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--ed04aa0b--4f5a--523b--a5d2--05072a1b260d-osd--block--ed04aa0b--4f5a--523b--a5d2--05072a1b260d', 'dm-uuid-LVM-viMZkgfP3jZUgTAo0bSRpd40fLtFkJJmO8XW5mmi1aRChVrq8X3DDkwGl8XRb5at'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--45b09425--beda--50b4--a83b--1fc50b5adca4-osd--block--45b09425--beda--50b4--a83b--1fc50b5adca4', 'dm-uuid-LVM-XzfTGfmLHoftdDBti3ot8l7sbRObO8yALj7bfZRysK7QQ9WjzIgwF2E0nO1Pc63W'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--ede169e1--d974--55ed--a938--c303cb89cb5b-osd--block--ede169e1--d974--55ed--a938--c303cb89cb5b', 'dm-uuid-LVM-bEBXWAqtQdfT6xKYbL0dF10We9CLdKgKh0pf0vpR2lQnL405kp4ndXs76m3o0J5Y'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a-osd--block--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a', 'dm-uuid-LVM-G1LwYmzPCvwLZRWX1y5oYSHlzZdIxlOhfbzzvktDUwfRIgHaiaw6r0MV8HVh1TLk'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part1', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part14', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part15', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part16', 'scsi-SQEMU_QEMU_HARDDISK_b6130fcd-394f-4d1f-ac75-5973103f7ce0-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--64dd100a--6bcc--5763--b557--0732c79bfd55-osd--block--64dd100a--6bcc--5763--b557--0732c79bfd55'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-HiPrG3-9qzo-WJZ0-1Smx-Bi1t-ReBz-6cdxta', 'scsi-0QEMU_QEMU_HARDDISK_f08311eb-e2ad-4eb7-8da9-ac6d899a2409', 'scsi-SQEMU_QEMU_HARDDISK_f08311eb-e2ad-4eb7-8da9-ac6d899a2409'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--5368e13a--d21f--5912--91bd--da537815d9e8-osd--block--5368e13a--d21f--5912--91bd--da537815d9e8'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-m0F2nF-j9UB-mZfT-bRv8-xW1c-EVch-ZfDvHt', 'scsi-0QEMU_QEMU_HARDDISK_e03b171b-b829-41a0-a0b6-cc46d85f013d', 'scsi-SQEMU_QEMU_HARDDISK_e03b171b-b829-41a0-a0b6-cc46d85f013d'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_9fa3c153-0e2a-45ef-8e48-e0095eb937ad', 'scsi-SQEMU_QEMU_HARDDISK_9fa3c153-0e2a-45ef-8e48-e0095eb937ad'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-21-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part1', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part14', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part15', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part16', 'scsi-SQEMU_QEMU_HARDDISK_0fb4a7c9-731d-4ebf-9945-e69275d38ed5-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part1', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part14', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part15', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part16', 'scsi-SQEMU_QEMU_HARDDISK_95d0155a-2e41-44f5-9fae-3beacda427fd-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--ed04aa0b--4f5a--523b--a5d2--05072a1b260d-osd--block--ed04aa0b--4f5a--523b--a5d2--05072a1b260d'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-nra7ru-O0w6-dM26-lz5o-2arZ-rXsE-lANagX', 'scsi-0QEMU_QEMU_HARDDISK_4e74a544-8b3a-4a05-98a1-b7cc13f13c18', 'scsi-SQEMU_QEMU_HARDDISK_4e74a544-8b3a-4a05-98a1-b7cc13f13c18'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--ede169e1--d974--55ed--a938--c303cb89cb5b-osd--block--ede169e1--d974--55ed--a938--c303cb89cb5b'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-HVhEyD-n8kM-PvHm-CUDD-tuRZ-igjC-Nzour2', 'scsi-0QEMU_QEMU_HARDDISK_f76ae8f0-65cc-4a62-b19c-23f8a40fbde6', 'scsi-SQEMU_QEMU_HARDDISK_f76ae8f0-65cc-4a62-b19c-23f8a40fbde6'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--45b09425--beda--50b4--a83b--1fc50b5adca4-osd--block--45b09425--beda--50b4--a83b--1fc50b5adca4'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-XQTe4H-QTjK-DLSs-Qjj6-RKjC-da4T-xXChra', 'scsi-0QEMU_QEMU_HARDDISK_666e73fe-87f5-4f65-9ff2-95a038c8b18d', 'scsi-SQEMU_QEMU_HARDDISK_666e73fe-87f5-4f65-9ff2-95a038c8b18d'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8ae4aabb-0174-4999-a1d0-4bf78cbb2965', 'scsi-SQEMU_QEMU_HARDDISK_8ae4aabb-0174-4999-a1d0-4bf78cbb2965'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a-osd--block--23a81e86--5308--55a4--b4d4--ad2cd0b09f4a'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-YifJdB-JYWE-HhRV-mF1g-0k83-hzWG-ctJZWy', 'scsi-0QEMU_QEMU_HARDDISK_d7bb81ba-6491-4b5a-bb82-f0b49314eac0', 'scsi-SQEMU_QEMU_HARDDISK_d7bb81ba-6491-4b5a-bb82-f0b49314eac0'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-14-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_cdffe5f1-7f0d-460e-a1db-d6846d226144', 'scsi-SQEMU_QEMU_HARDDISK_cdffe5f1-7f0d-460e-a1db-d6846d226144'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-21-00-03-28-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Saturday 21 March 2026  00:56:05 +0000 (0:00:00.527)       0:00:16.809 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Saturday 21 March 2026  00:56:06 +0000 (0:00:00.560)       0:00:17.370 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Saturday 21 March 2026  00:56:06 +0000 (0:00:00.413)       0:00:17.784 ******** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Saturday 21 March 2026  00:56:07 +0000 (0:00:01.430)       0:00:19.214 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Saturday 21 March 2026  00:56:08 +0000 (0:00:00.238)       0:00:19.453 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Saturday 21 March 2026  00:56:08 +0000 (0:00:00.390)       0:00:19.844 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Saturday 21 March 2026  00:56:08 +0000 (0:00:00.392)       0:00:20.236 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Saturday 21 March 2026  00:56:09 +0000 (0:00:00.967)       0:00:21.203 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Saturday 21 March 2026  00:56:10 +0000 (0:00:00.301)       0:00:21.504 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Saturday 21 March 2026  00:56:10 +0000 (0:00:00.584)       0:00:22.088 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Saturday 21 March 2026  00:56:11 +0000 (0:00:00.260)       0:00:22.349 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Saturday 21 March 2026  00:56:11 +0000 (0:00:00.273)       0:00:22.622 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Saturday 21 March 2026  00:56:11 +0000 (0:00:00.256)       0:00:22.879 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Saturday 21 March 2026  00:56:12 +0000 (0:00:00.464)       0:00:23.344 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Saturday 21 March 2026  00:56:12 +0000 (0:00:00.343)       0:00:23.687 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Saturday 21 March 2026  00:56:12 +0000 (0:00:00.312)       0:00:24.000 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Saturday 21 March 2026  00:56:13 +0000 (0:00:00.314)       0:00:24.314 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Saturday 21 March 2026  00:56:13 +0000 (0:00:00.278)       0:00:24.593 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Saturday 21 March 2026  00:56:13 +0000 (0:00:00.531)       0:00:25.125 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Saturday 21 March 2026  00:56:14 +0000 (0:00:01.061)       0:00:26.186 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [Include tasks from the ceph-osd role] ************************************",
                                "Saturday 21 March 2026  00:56:17 +0000 (0:00:02.256)       0:00:28.443 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/openstack_config.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [create openstack pool(s)] ************************************************",
                                "Saturday 21 March 2026  00:56:17 +0000 (0:00:00.485)       0:00:28.929 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'backups', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'volumes', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'images', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'metrics', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'vms', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "",
                                "TASK [generate keys] ***********************************************************",
                                "Saturday 21 March 2026  00:57:02 +0000 (0:00:45.248)       0:01:14.178 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> {{ groups[mon_group_name][0] }}]\u001b[0m",
                                "",
                                "TASK [get keys from monitors] **************************************************",
                                "Saturday 21 March 2026  00:57:24 +0000 (0:00:21.264)       0:01:35.442 ******** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [copy ceph key(s) if needed] **********************************************",
                                "Saturday 21 March 2026  00:57:36 +0000 (0:00:12.491)       0:01:47.934 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> {{ item.1 }}]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=25  \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=28  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=21  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:57:54 +0000 (0:00:17.558)       0:02:05.492 ******** ",
                                "=============================================================================== ",
                                "create openstack pool(s) ----------------------------------------------- 45.25s",
                                "generate keys ---------------------------------------------------------- 21.26s",
                                "copy ceph key(s) if needed --------------------------------------------- 17.56s",
                                "get keys from monitors ------------------------------------------------- 12.49s",
                                "ceph-facts : Find a running mon container ------------------------------- 2.79s",
                                "ceph-facts : Set_fact ceph_admin_command -------------------------------- 2.26s",
                                "ceph-facts : Get current fsid if cluster is already running ------------- 1.60s",
                                "ceph-facts : Read osd pool default crush rule --------------------------- 1.43s",
                                "ceph-facts : Set_fact ceph_run_cmd -------------------------------------- 1.06s",
                                "ceph-facts : Set_fact _monitor_addresses - ipv4 ------------------------- 0.97s",
                                "ceph-facts : Check if it is atomic host --------------------------------- 0.84s",
                                "ceph-facts : Check if the ceph mon socket is in-use --------------------- 0.84s",
                                "ceph-facts : Check if podman binary is present -------------------------- 0.68s",
                                "ceph-facts : Set_fact monitor_name ansible_facts['hostname'] ------------ 0.64s",
                                "ceph-facts : Import_tasks set_radosgw_address.yml ----------------------- 0.58s",
                                "ceph-facts : Check if the ceph conf exists ------------------------------ 0.56s",
                                "ceph-facts : Set_fact _container_exec_cmd ------------------------------- 0.55s",
                                "ceph-facts : Collect existed devices ------------------------------------ 0.55s",
                                "ceph-facts : Set_fact rgw_instances ------------------------------------- 0.53s",
                                "ceph-facts : Set_fact devices generate device list when osd_auto_discovery --- 0.53s",
                                "\u001b[32m2026-03-21 00:57:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:57:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 20a919ac-e868-47dc-b863-2b49bac96899 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8ff0fa64-c4d0-4a36-af2c-ad3fc6ccdff9 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Copy ceph keys to the configuration repository] **************************",
                                "",
                                "TASK [Check if ceph keys exist] ************************************************",
                                "Saturday 21 March 2026  00:57:57 +0000 (0:00:00.237)       0:00:00.237 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.admin.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.nova.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.glance.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.gnocchi.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.manila.keyring)\u001b[0m",
                                "",
                                "TASK [Fetch all ceph keys] *****************************************************",
                                "Saturday 21 March 2026  00:58:02 +0000 (0:00:04.863)       0:00:05.101 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.admin.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.nova.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.glance.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.gnocchi.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.manila.keyring)\u001b[0m",
                                "",
                                "TASK [Create share directory] **************************************************",
                                "Saturday 21 March 2026  00:58:06 +0000 (0:00:04.223)       0:00:09.324 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Write ceph keys to the share directory] **********************************",
                                "Saturday 21 March 2026  00:58:08 +0000 (0:00:01.055)       0:00:10.379 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.admin.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.nova.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.glance.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.gnocchi.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.manila.keyring)\u001b[0m",
                                "",
                                "TASK [Check if target directories exist] ***************************************",
                                "Saturday 21 March 2026  00:58:21 +0000 (0:00:13.159)       0:00:23.539 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/infrastructure/files/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-volume)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-backup)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-backup)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/nova)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/nova)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/glance)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/gnocchi)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/manila)\u001b[0m",
                                "",
                                "TASK [Write ceph keys to the configuration directory] **************************",
                                "Saturday 21 March 2026  00:58:24 +0000 (0:00:03.333)       0:00:26.872 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.admin.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.nova.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.glance.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.gnocchi.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.manila.keyring)\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=6   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:58:31 +0000 (0:00:06.889)       0:00:33.761 ******** ",
                                "=============================================================================== ",
                                "Write ceph keys to the share directory --------------------------------- 13.16s",
                                "Write ceph keys to the configuration directory -------------------------- 6.89s",
                                "Check if ceph keys exist ------------------------------------------------ 4.86s",
                                "Fetch all ceph keys ----------------------------------------------------- 4.22s",
                                "Check if target directories exist --------------------------------------- 3.33s",
                                "Create share directory -------------------------------------------------- 1.06s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  00:56:03 +0000 (0:00:00.279)       0:00:00.279 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  00:56:04 +0000 (0:00:00.259)       0:00:00.539 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_keystone_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_keystone_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_keystone_True)\u001b[0m",
                                "",
                                "PLAY [Apply role keystone] *****************************************************",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Saturday 21 March 2026  00:56:04 +0000 (0:00:00.274)       0:00:00.813 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/keystone/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [keystone : Ensuring config directories exist] ****************************",
                                "Saturday 21 March 2026  00:56:04 +0000 (0:00:00.608)       0:00:01.422 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [keystone : Check if policies shall be overwritten] ***********************",
                                "Saturday 21 March 2026  00:56:07 +0000 (0:00:02.250)       0:00:03.673 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Set keystone policy file] *************************************",
                                "Saturday 21 March 2026  00:56:07 +0000 (0:00:00.118)       0:00:03.792 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Check if Keystone domain-specific config is supplied] *********",
                                "Saturday 21 March 2026  00:56:07 +0000 (0:00:00.247)       0:00:04.039 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Saturday 21 March 2026  00:56:08 +0000 (0:00:00.831)       0:00:04.871 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/keystone/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : keystone | Copying over extra CA certificates] *******",
                                "Saturday 21 March 2026  00:56:09 +0000 (0:00:00.583)       0:00:05.455 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : keystone | Copying over backend internal TLS certificate] ***",
                                "Saturday 21 March 2026  00:56:11 +0000 (0:00:02.937)       0:00:08.392 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : keystone | Copying over backend internal TLS key] ****",
                                "Saturday 21 March 2026  00:56:12 +0000 (0:00:00.516)       0:00:08.908 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying over config.json files for services] ******************",
                                "Saturday 21 March 2026  00:56:13 +0000 (0:00:00.893)       0:00:09.802 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [keystone : Copying over keystone.conf] ***********************************",
                                "Saturday 21 March 2026  00:56:17 +0000 (0:00:03.814)       0:00:13.617 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [keystone : Copying keystone-startup script for keystone] *****************",
                                "Saturday 21 March 2026  00:56:22 +0000 (0:00:05.359)       0:00:18.977 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Create Keystone domain-specific config directory] *************",
                                "Saturday 21 March 2026  00:56:24 +0000 (0:00:01.605)       0:00:20.582 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Get file list in custom domains folder] ***********************",
                                "Saturday 21 March 2026  00:56:25 +0000 (0:00:00.959)       0:00:21.542 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying Keystone Domain specific settings] ********************",
                                "Saturday 21 March 2026  00:56:25 +0000 (0:00:00.300)       0:00:21.843 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying over existing policy file] ****************************",
                                "Saturday 21 March 2026  00:56:25 +0000 (0:00:00.299)       0:00:22.143 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Saturday 21 March 2026  00:56:26 +0000 (0:00:00.561)       0:00:22.705 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying over wsgi-keystone.conf] ******************************",
                                "Saturday 21 March 2026  00:56:26 +0000 (0:00:00.446)       0:00:23.152 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m",
                                "",
                                "TASK [keystone : Checking whether keystone-paste.ini file exists] **************",
                                "Saturday 21 March 2026  00:56:28 +0000 (0:00:01.587)       0:00:24.740 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [keystone : Copying over keystone-paste.ini] ******************************",
                                "Saturday 21 March 2026  00:56:29 +0000 (0:00:00.971)       0:00:25.711 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Generate the required cron jobs for the node] *****************",
                                "Saturday 21 March 2026  00:56:29 +0000 (0:00:00.488)       0:00:26.199 ******** ",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "",
                                "TASK [keystone : Set fact with the generated cron jobs for building the crontab later] ***",
                                "Saturday 21 March 2026  00:56:30 +0000 (0:00:01.120)       0:00:27.320 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying files for keystone-fernet] ****************************",
                                "Saturday 21 March 2026  00:56:31 +0000 (0:00:00.435)       0:00:27.756 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "",
                                "TASK [keystone : Copying files for keystone-ssh] *******************************",
                                "Saturday 21 March 2026  00:56:39 +0000 (0:00:08.262)       0:00:36.018 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "",
                                "TASK [keystone : Check keystone containers] ************************************",
                                "Saturday 21 March 2026  00:56:42 +0000 (0:00:02.496)       0:00:38.514 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Saturday 21 March 2026  00:56:44 +0000 (0:00:02.072)       0:00:40.586 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Creating keystone database] ***********************************",
                                "Saturday 21 March 2026  00:56:44 +0000 (0:00:00.277)       0:00:40.864 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Creating Keystone database user and setting permissions] ******",
                                "Saturday 21 March 2026  00:56:47 +0000 (0:00:02.939)       0:00:43.803 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Checking for any running keystone_fernet containers] **********",
                                "Saturday 21 March 2026  00:56:49 +0000 (0:00:02.488)       0:00:46.292 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Group nodes where keystone_fernet is running] *****************",
                                "Saturday 21 March 2026  00:56:50 +0000 (0:00:00.787)       0:00:47.080 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Fail if any hosts need bootstrapping and not all hosts targeted] ***",
                                "Saturday 21 March 2026  00:56:50 +0000 (0:00:00.315)       0:00:47.395 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Running Keystone bootstrap container] *************************",
                                "Saturday 21 March 2026  00:56:51 +0000 (0:00:00.312)       0:00:47.708 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Running Keystone fernet bootstrap container] ******************",
                                "Saturday 21 March 2026  00:57:05 +0000 (0:00:14.282)       0:01:01.990 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Flush handlers] ***********************************************",
                                "Saturday 21 March 2026  00:57:16 +0000 (0:00:10.938)       0:01:12.929 ******** ",
                                "",
                                "TASK [keystone : Flush handlers] ***********************************************",
                                "Saturday 21 March 2026  00:57:16 +0000 (0:00:00.062)       0:01:12.992 ******** ",
                                "",
                                "TASK [keystone : Flush handlers] ***********************************************",
                                "Saturday 21 March 2026  00:57:16 +0000 (0:00:00.064)       0:01:13.056 ******** ",
                                "",
                                "RUNNING HANDLER [keystone : Restart keystone-ssh container] ********************",
                                "Saturday 21 March 2026  00:57:16 +0000 (0:00:00.067)       0:01:13.123 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [keystone : Restart keystone-fernet container] *****************",
                                "Saturday 21 March 2026  00:57:26 +0000 (0:00:10.084)       0:01:23.208 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [keystone : Restart keystone container] ************************",
                                "Saturday 21 March 2026  00:57:36 +0000 (0:00:09.863)       0:01:33.071 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Saturday 21 March 2026  00:57:42 +0000 (0:00:05.760)       0:01:38.832 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/keystone/tasks/distribute_fernet.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [keystone : Waiting for Keystone SSH port to be UP] ***********************",
                                "Saturday 21 March 2026  00:57:43 +0000 (0:00:00.677)       0:01:39.510 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [keystone : Run key distribution] *****************************************",
                                "Saturday 21 March 2026  00:57:43 +0000 (0:00:00.746)       0:01:40.257 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Creating admin project, user, role, service, and endpoint] ****",
                                "Saturday 21 March 2026  00:57:45 +0000 (0:00:01.740)       0:01:41.998 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=RegionOne)\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating services] **********************",
                                "Saturday 21 March 2026  00:57:58 +0000 (0:00:12.512)       0:01:54.510 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=keystone (identity))\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating endpoints] *********************",
                                "Saturday 21 March 2026  00:58:27 +0000 (0:00:29.143)       0:02:23.653 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=keystone -> https://api-int.testbed.osism.xyz:5000 -> internal)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=keystone -> https://api.testbed.osism.xyz:5000 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating projects] **********************",
                                "Saturday 21 March 2026  00:58:34 +0000 (0:00:07.179)       0:02:30.833 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating users] *************************",
                                "Saturday 21 March 2026  00:58:34 +0000 (0:00:00.119)       0:02:30.953 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating roles] *************************",
                                "Saturday 21 March 2026  00:58:34 +0000 (0:00:00.114)       0:02:31.068 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Granting user roles] ********************",
                                "Saturday 21 March 2026  00:58:34 +0000 (0:00:00.119)       0:02:31.187 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Creating default user role] ***********************************",
                                "Saturday 21 March 2026  00:58:35 +0000 (0:00:00.313)       0:02:31.501 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Saturday 21 March 2026  00:58:38 +0000 (0:00:03.228)       0:02:34.730 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=33  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:58:38 +0000 (0:00:00.551)       0:02:35.281 ******** ",
                                "=============================================================================== ",
                                "service-ks-register : keystone | Creating services --------------------- 29.14s",
                                "keystone : Running Keystone bootstrap container ------------------------ 14.28s",
                                "keystone : Creating admin project, user, role, service, and endpoint --- 12.51s",
                                "keystone : Running Keystone fernet bootstrap container ----------------- 10.94s",
                                "keystone : Restart keystone-ssh container ------------------------------ 10.08s",
                                "keystone : Restart keystone-fernet container ---------------------------- 9.86s",
                                "keystone : Copying files for keystone-fernet ---------------------------- 8.26s",
                                "service-ks-register : keystone | Creating endpoints --------------------- 7.18s",
                                "keystone : Restart keystone container ----------------------------------- 5.76s",
                                "keystone : Copying over keystone.conf ----------------------------------- 5.36s",
                                "keystone : Copying over config.json files for services ------------------ 3.81s",
                                "keystone : Creating default user role ----------------------------------- 3.23s",
                                "keystone : Creating keystone database ----------------------------------- 2.94s",
                                "service-cert-copy : keystone | Copying over extra CA certificates ------- 2.94s",
                                "keystone : Copying files for keystone-ssh ------------------------------- 2.50s",
                                "keystone : Creating Keystone database user and setting permissions ------ 2.49s",
                                "keystone : Ensuring config directories exist ---------------------------- 2.25s",
                                "keystone : Check keystone containers ------------------------------------ 2.07s",
                                "keystone : Run key distribution ----------------------------------------- 1.74s",
                                "keystone : Copying keystone-startup script for keystone ----------------- 1.61s",
                                "\u001b[32m2026-03-21 00:58:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:58:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f6b4f1d9-f2a4-4a9e-95ff-71c06aa23c12 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 00:59:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c43f457-da25-4332-acf5-2db48ac793bf is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Apply role cephclient] ***************************************************",
                                "",
                                "TASK [osism.services.cephclient : Include container tasks] *********************",
                                "Saturday 21 March 2026  00:58:35 +0000 (0:00:00.306)       0:00:00.306 ******** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/cephclient/tasks/container.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Create required directories] *****************",
                                "Saturday 21 March 2026  00:58:35 +0000 (0:00:00.241)       0:00:00.547 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/cephclient/configuration)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/cephclient/data)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/cephclient)\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Copy configuration files] ********************",
                                "Saturday 21 March 2026  00:58:36 +0000 (0:00:01.557)       0:00:02.105 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'src': 'ceph.conf.j2', 'dest': '/opt/cephclient/configuration/ceph.conf'})\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Copy keyring file] ***************************",
                                "Saturday 21 March 2026  00:58:38 +0000 (0:00:01.190)       0:00:03.295 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Copy docker-compose.yml file] ****************",
                                "Saturday 21 March 2026  00:58:38 +0000 (0:00:00.867)       0:00:04.163 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Manage cephclient service] *******************",
                                "Saturday 21 March 2026  00:58:39 +0000 (0:00:00.898)       0:00:05.061 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage cephclient service (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Copy wrapper scripts] ************************",
                                "Saturday 21 March 2026  00:59:15 +0000 (0:00:35.560)       0:00:40.622 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph-authtool)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=rados)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=radosgw-admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=rbd)\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Remove old wrapper scripts] ******************",
                                "Saturday 21 March 2026  00:59:18 +0000 (0:00:03.550)       0:00:44.173 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=crushtool)\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Include package tasks] ***********************",
                                "Saturday 21 March 2026  00:59:19 +0000 (0:00:00.514)       0:00:44.687 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Include rook task] ***************************",
                                "Saturday 21 March 2026  00:59:19 +0000 (0:00:00.108)       0:00:44.795 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.cephclient : Restart cephclient service] *******",
                                "Saturday 21 March 2026  00:59:19 +0000 (0:00:00.269)       0:00:45.065 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.cephclient : Ensure that all containers are up] ***",
                                "Saturday 21 March 2026  00:59:21 +0000 (0:00:01.333)       0:00:46.398 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.cephclient : Wait for an healthy service] ******",
                                "Saturday 21 March 2026  00:59:21 +0000 (0:00:00.809)       0:00:47.207 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.cephclient : Copy bash completion scripts] *****",
                                "Saturday 21 March 2026  00:59:22 +0000 (0:00:00.527)       0:00:47.735 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=rados)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=radosgw-admin)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=rbd)\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  00:59:23 +0000 (0:00:01.142)       0:00:48.877 ******** ",
                                "=============================================================================== ",
                                "osism.services.cephclient : Manage cephclient service ------------------ 35.56s",
                                "osism.services.cephclient : Copy wrapper scripts ------------------------ 3.55s",
                                "osism.services.cephclient : Create required directories ----------------- 1.56s",
                                "osism.services.cephclient : Restart cephclient service ------------------ 1.33s",
                                "osism.services.cephclient : Copy configuration files -------------------- 1.19s",
                                "osism.services.cephclient : Copy bash completion scripts ---------------- 1.14s",
                                "osism.services.cephclient : Copy docker-compose.yml file ---------------- 0.90s",
                                "osism.services.cephclient : Copy keyring file --------------------------- 0.87s",
                                "osism.services.cephclient : Ensure that all containers are up ----------- 0.81s",
                                "osism.services.cephclient : Wait for an healthy service ----------------- 0.53s",
                                "osism.services.cephclient : Remove old wrapper scripts ------------------ 0.51s",
                                "osism.services.cephclient : Include rook task --------------------------- 0.27s",
                                "osism.services.cephclient : Include container tasks --------------------- 0.24s",
                                "osism.services.cephclient : Include package tasks ----------------------- 0.11s",
                                "",
                                "",
                                "PLAY [Download ironic ipa images] **********************************************",
                                "",
                                "TASK [Ensure the destination directory exists] *********************************",
                                "Saturday 21 March 2026  00:58:43 +0000 (0:00:00.175)       0:00:00.175 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Download ironic-agent initramfs] *****************************************",
                                "Saturday 21 March 2026  00:58:45 +0000 (0:00:01.423)       0:00:01.598 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Download ironic-agent initramfs (3 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Download ironic-agent initramfs (2 retries left).\u001b[0m",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Download ironic-agent kernel] ********************************************",
                                "Saturday 21 March 2026  01:00:05 +0000 (0:01:20.183)       0:01:21.782 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  01:00:11 +0000 (0:00:06.172)       0:01:27.955 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  01:00:11 +0000 (0:00:00.302)       0:01:28.257 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_ironic_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_ironic_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_ironic_False)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: enable_ironic_True\u001b[0m",
                                "",
                                "PLAY [Apply role ironic] *******************************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  01:00:12 +0000 (0:00:00.376)       0:01:28.634 ******** ",
                                "=============================================================================== ",
                                "Download ironic-agent initramfs ---------------------------------------- 80.18s",
                                "Download ironic-agent kernel -------------------------------------------- 6.17s",
                                "Ensure the destination directory exists --------------------------------- 1.42s",
                                "Group hosts based on enabled services ----------------------------------- 0.38s",
                                "Group hosts based on Kolla action --------------------------------------- 0.30s",
                                "\u001b[32m2026-03-21 01:00:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 079d9c76-84a6-42c0-ace7-bef8512ca035 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  00:58:43 +0000 (0:00:00.458)       0:00:00.458 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  00:58:43 +0000 (0:00:00.393)       0:00:00.851 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_barbican_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_barbican_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_barbican_True)\u001b[0m",
                                "",
                                "PLAY [Apply role barbican] *****************************************************",
                                "",
                                "TASK [barbican : include_tasks] ************************************************",
                                "Saturday 21 March 2026  00:58:44 +0000 (0:00:00.375)       0:00:01.227 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/barbican/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating services] **********************",
                                "Saturday 21 March 2026  00:58:45 +0000 (0:00:00.750)       0:00:01.977 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican (key-manager))\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating endpoints] *********************",
                                "Saturday 21 March 2026  00:58:49 +0000 (0:00:04.135)       0:00:06.113 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> https://api-int.testbed.osism.xyz:9311 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> https://api.testbed.osism.xyz:9311 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating projects] **********************",
                                "Saturday 21 March 2026  00:58:56 +0000 (0:00:07.062)       0:00:13.176 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating users] *************************",
                                "Saturday 21 March 2026  00:58:59 +0000 (0:00:03.627)       0:00:16.803 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating roles] *************************",
                                "Saturday 21 March 2026  00:59:03 +0000 (0:00:03.738)       0:00:20.541 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=key-manager:service-admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=creator)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=observer)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=audit)\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Granting user roles] ********************",
                                "Saturday 21 March 2026  00:59:20 +0000 (0:00:17.225)       0:00:37.767 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> service -> admin)\u001b[0m",
                                "",
                                "TASK [barbican : Ensuring config directories exist] ****************************",
                                "Saturday 21 March 2026  00:59:25 +0000 (0:00:04.563)       0:00:42.331 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [barbican : Ensuring vassals config directories exist] ********************",
                                "Saturday 21 March 2026  00:59:28 +0000 (0:00:02.996)       0:00:45.327 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican-api/vassals)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=barbican-api/vassals)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=barbican-api/vassals)\u001b[0m",
                                "",
                                "TASK [barbican : Check if policies shall be overwritten] ***********************",
                                "Saturday 21 March 2026  00:59:29 +0000 (0:00:01.443)       0:00:46.770 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [barbican : Set barbican policy file] *************************************",
                                "Saturday 21 March 2026  00:59:30 +0000 (0:00:00.157)       0:00:46.928 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : include_tasks] ************************************************",
                                "Saturday 21 March 2026  00:59:30 +0000 (0:00:00.236)       0:00:47.165 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/barbican/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : barbican | Copying over extra CA certificates] *******",
                                "Saturday 21 March 2026  00:59:30 +0000 (0:00:00.527)       0:00:47.693 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : barbican | Copying over backend internal TLS certificate] ***",
                                "Saturday 21 March 2026  00:59:35 +0000 (0:00:04.292)       0:00:51.985 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : barbican | Copying over backend internal TLS key] ****",
                                "Saturday 21 March 2026  00:59:36 +0000 (0:00:01.296)       0:00:53.282 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [barbican : Copying over config.json files for services] ******************",
                                "Saturday 21 March 2026  00:59:37 +0000 (0:00:00.759)       0:00:54.041 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [barbican : Copying over barbican-api.ini] ********************************",
                                "Saturday 21 March 2026  00:59:42 +0000 (0:00:05.318)       0:00:59.360 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : Checking whether barbican-api-paste.ini file exists] **********",
                                "Saturday 21 March 2026  00:59:44 +0000 (0:00:01.844)       0:01:01.204 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [barbican : Copying over barbican-api-paste.ini] **************************",
                                "Saturday 21 March 2026  00:59:46 +0000 (0:00:02.145)       0:01:03.350 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : Copying over barbican.conf] ***********************************",
                                "Saturday 21 March 2026  00:59:47 +0000 (0:00:00.661)       0:01:04.012 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [barbican : Copying over existing policy file] ****************************",
                                "Saturday 21 March 2026  00:59:58 +0000 (0:00:11.290)       0:01:15.302 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : Check barbican containers] ************************************",
                                "Saturday 21 March 2026  01:00:00 +0000 (0:00:01.932)       0:01:17.236 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [barbican : include_tasks] ************************************************",
                                "Saturday 21 March 2026  01:00:03 +0000 (0:00:03.462)       0:01:20.699 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : Creating barbican database] ***********************************",
                                "Saturday 21 March 2026  01:00:04 +0000 (0:00:00.813)       0:01:21.512 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [barbican : Creating barbican database user and setting permissions] ******",
                                "Saturday 21 March 2026  01:00:07 +0000 (0:00:02.377)       0:01:23.889 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [barbican : Running barbican bootstrap container] *************************",
                                "Saturday 21 March 2026  01:00:09 +0000 (0:00:02.365)       0:01:26.254 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [barbican : Flush handlers] ***********************************************",
                                "Saturday 21 March 2026  01:00:20 +0000 (0:00:11.433)       0:01:37.688 ******** ",
                                "",
                                "TASK [barbican : Flush handlers] ***********************************************",
                                "Saturday 21 March 2026  01:00:21 +0000 (0:00:00.339)       0:01:38.028 ******** ",
                                "",
                                "TASK [barbican : Flush handlers] ***********************************************",
                                "Saturday 21 March 2026  01:00:21 +0000 (0:00:00.068)       0:01:38.096 ******** ",
                                "",
                                "RUNNING HANDLER [barbican : Restart barbican-api container] ********************",
                                "Saturday 21 March 2026  01:00:21 +0000 (0:00:00.065)       0:01:38.162 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [barbican : Restart barbican-keystone-listener container] ******",
                                "Saturday 21 March 2026  01:00:31 +0000 (0:00:10.300)       0:01:48.462 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [barbican : Restart barbican-worker container] *****************",
                                "Saturday 21 March 2026  01:00:41 +0000 (0:00:10.064)       0:01:58.527 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  01:00:52 +0000 (0:00:10.839)       0:02:09.367 ******** ",
                                "=============================================================================== ",
                                "service-ks-register : barbican | Creating roles ------------------------ 17.23s",
                                "barbican : Running barbican bootstrap container ------------------------ 11.43s",
                                "barbican : Copying over barbican.conf ---------------------------------- 11.29s",
                                "barbican : Restart barbican-worker container --------------------------- 10.84s",
                                "barbican : Restart barbican-api container ------------------------------ 10.30s",
                                "barbican : Restart barbican-keystone-listener container ---------------- 10.06s",
                                "service-ks-register : barbican | Creating endpoints --------------------- 7.06s",
                                "barbican : Copying over config.json files for services ------------------ 5.32s",
                                "service-ks-register : barbican | Granting user roles -------------------- 4.56s",
                                "service-cert-copy : barbican | Copying over extra CA certificates ------- 4.29s",
                                "service-ks-register : barbican | Creating services ---------------------- 4.14s",
                                "service-ks-register : barbican | Creating users ------------------------- 3.74s",
                                "service-ks-register : barbican | Creating projects ---------------------- 3.63s",
                                "barbican : Check barbican containers ------------------------------------ 3.46s",
                                "barbican : Ensuring config directories exist ---------------------------- 3.00s",
                                "barbican : Creating barbican database ----------------------------------- 2.38s",
                                "barbican : Creating barbican database user and setting permissions ------ 2.37s",
                                "barbican : Checking whether barbican-api-paste.ini file exists ---------- 2.15s",
                                "barbican : Copying over existing policy file ---------------------------- 1.93s",
                                "barbican : Copying over barbican-api.ini -------------------------------- 1.84s",
                                "\u001b[32m2026-03-21 01:00:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:00:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff0575d-40a3-464c-b102-19fbdc0c5555 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5ad1be9d-918f-439e-8fe3-c7925cb9fc3d is in state SUCCESS\u001b[0m",
                                "",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Bootstraph ceph dashboard] ***********************************************",
                                "",
                                "TASK [Disable the ceph dashboard] **********************************************",
                                "Saturday 21 March 2026  00:59:27 +0000 (0:00:00.191)       0:00:00.191 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/ssl to false] ******************************************",
                                "Saturday 21 March 2026  00:59:30 +0000 (0:00:02.428)       0:00:02.619 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/server_port to 7000] ***********************************",
                                "Saturday 21 March 2026  00:59:31 +0000 (0:00:00.908)       0:00:03.527 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/server_addr to 0.0.0.0] ********************************",
                                "Saturday 21 March 2026  00:59:31 +0000 (0:00:00.842)       0:00:04.370 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/standby_behaviour to error] ****************************",
                                "Saturday 21 March 2026  00:59:32 +0000 (0:00:01.064)       0:00:05.435 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/standby_error_status_code to 404] **********************",
                                "Saturday 21 March 2026  00:59:34 +0000 (0:00:01.278)       0:00:06.713 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Enable the ceph dashboard] ***********************************************",
                                "Saturday 21 March 2026  00:59:35 +0000 (0:00:01.151)       0:00:07.865 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Write ceph_dashboard_password to temporary file] *************************",
                                "Saturday 21 March 2026  00:59:37 +0000 (0:00:02.012)       0:00:09.877 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Create admin user] *******************************************************",
                                "Saturday 21 March 2026  00:59:38 +0000 (0:00:01.156)       0:00:11.033 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Remove temporary file for ceph_dashboard_password] ***********************",
                                "Saturday 21 March 2026  01:00:35 +0000 (0:00:56.532)       0:01:07.566 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Restart ceph manager services] *******************************************",
                                "",
                                "TASK [Restart ceph manager service] ********************************************",
                                "Saturday 21 March 2026  01:00:35 +0000 (0:00:00.133)       0:01:07.699 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Restart ceph manager services] *******************************************",
                                "",
                                "TASK [Restart ceph manager service] ********************************************",
                                "Saturday 21 March 2026  01:00:46 +0000 (0:00:11.530)       0:01:19.229 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Restart ceph manager services] *******************************************",
                                "",
                                "TASK [Restart ceph manager service] ********************************************",
                                "Saturday 21 March 2026  01:00:48 +0000 (0:00:01.299)       0:01:20.529 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  01:00:59 +0000 (0:00:11.323)       0:01:31.852 ******** ",
                                "=============================================================================== ",
                                "Create admin user ------------------------------------------------------ 56.53s",
                                "Restart ceph manager service ------------------------------------------- 24.15s",
                                "Disable the ceph dashboard ---------------------------------------------- 2.43s",
                                "Enable the ceph dashboard ----------------------------------------------- 2.01s",
                                "Set mgr/dashboard/standby_behaviour to error ---------------------------- 1.28s",
                                "Write ceph_dashboard_password to temporary file ------------------------- 1.16s",
                                "Set mgr/dashboard/standby_error_status_code to 404 ---------------------- 1.15s",
                                "Set mgr/dashboard/server_addr to 0.0.0.0 -------------------------------- 1.06s",
                                "Set mgr/dashboard/ssl to false ------------------------------------------ 0.91s",
                                "Set mgr/dashboard/server_port to 7000 ----------------------------------- 0.84s",
                                "Remove temporary file for ceph_dashboard_password ----------------------- 0.13s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  01:00:17 +0000 (0:00:00.413)       0:00:00.413 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  01:00:17 +0000 (0:00:00.256)       0:00:00.669 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_placement_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_placement_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_placement_True)\u001b[0m",
                                "",
                                "PLAY [Apply role placement] ****************************************************",
                                "",
                                "TASK [placement : include_tasks] ***********************************************",
                                "Saturday 21 March 2026  01:00:18 +0000 (0:00:00.276)       0:00:00.945 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/placement/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating services] *********************",
                                "Saturday 21 March 2026  01:00:19 +0000 (0:00:01.016)       0:00:01.962 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement (placement))\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating endpoints] ********************",
                                "Saturday 21 March 2026  01:00:22 +0000 (0:00:03.374)       0:00:05.336 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> https://api-int.testbed.osism.xyz:8780 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> https://api.testbed.osism.xyz:8780 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating projects] *********************",
                                "Saturday 21 March 2026  01:00:28 +0000 (0:00:05.813)       0:00:11.150 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating users] ************************",
                                "Saturday 21 March 2026  01:00:32 +0000 (0:00:03.763)       0:00:14.914 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating roles] ************************",
                                "Saturday 21 March 2026  01:00:35 +0000 (0:00:03.720)       0:00:18.634 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Granting user roles] *******************",
                                "Saturday 21 March 2026  01:00:38 +0000 (0:00:03.049)       0:00:21.684 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> service -> admin)\u001b[0m",
                                "",
                                "TASK [placement : include_tasks] ***********************************************",
                                "Saturday 21 March 2026  01:00:43 +0000 (0:00:04.235)       0:00:25.919 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [placement : Ensuring config directories exist] ***************************",
                                "Saturday 21 March 2026  01:00:43 +0000 (0:00:00.527)       0:00:26.447 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [placement : Check if policies shall be overwritten] **********************",
                                "Saturday 21 March 2026  01:00:45 +0000 (0:00:02.113)       0:00:28.560 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [placement : Set placement policy file] ***********************************",
                                "Saturday 21 March 2026  01:00:45 +0000 (0:00:00.124)       0:00:28.685 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [placement : include_tasks] ***********************************************",
                                "Saturday 21 March 2026  01:00:46 +0000 (0:00:00.254)       0:00:28.939 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/placement/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : placement | Copying over extra CA certificates] ******",
                                "Saturday 21 March 2026  01:00:47 +0000 (0:00:01.085)       0:00:30.025 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : placement | Copying over backend internal TLS certificate] ***",
                                "Saturday 21 March 2026  01:00:49 +0000 (0:00:01.991)       0:00:32.016 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : placement | Copying over backend internal TLS key] ***",
                                "Saturday 21 March 2026  01:00:50 +0000 (0:00:01.008)       0:00:33.024 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [placement : Copying over config.json files for services] *****************",
                                "Saturday 21 March 2026  01:00:51 +0000 (0:00:01.297)       0:00:34.322 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [placement : Copying over placement.conf] *********************************",
                                "Saturday 21 March 2026  01:00:53 +0000 (0:00:01.735)       0:00:36.058 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [placement : Copying over placement-api wsgi configuration] ***************",
                                "Saturday 21 March 2026  01:00:55 +0000 (0:00:02.449)       0:00:38.508 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2)\u001b[0m",
                                "",
                                "TASK [placement : Copying over migrate-db.rc.j2 configuration] *****************",
                                "Saturday 21 March 2026  01:00:57 +0000 (0:00:01.622)       0:00:40.130 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [placement : Copying over existing policy file] ***************************",
                                "Saturday 21 March 2026  01:00:58 +0000 (0:00:01.502)       0:00:41.632 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [placement : Check placement containers] **********************************",
                                "Saturday 21 March 2026  01:01:00 +0000 (0:00:01.687)       0:00:43.319 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [placement : Creating placement databases] ********************************",
                                "Saturday 21 March 2026  01:01:02 +0000 (0:00:01.461)       0:00:44.780 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [placement : Creating placement databases user and setting permissions] ***",
                                "Saturday 21 March 2026  01:01:03 +0000 (0:00:01.960)       0:00:46.741 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [placement : Running placement bootstrap container] ***********************",
                                "Saturday 21 March 2026  01:01:06 +0000 (0:00:02.140)       0:00:48.881 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [placement : Flush handlers] **********************************************",
                                "Saturday 21 March 2026  01:01:18 +0000 (0:00:12.443)       0:01:01.325 ******** ",
                                "",
                                "TASK [placement : Flush handlers] **********************************************",
                                "Saturday 21 March 2026  01:01:18 +0000 (0:00:00.056)       0:01:01.381 ******** ",
                                "",
                                "TASK [placement : Flush handlers] **********************************************",
                                "Saturday 21 March 2026  01:01:18 +0000 (0:00:00.059)       0:01:01.441 ******** ",
                                "",
                                "RUNNING HANDLER [placement : Restart placement-api container] ******************",
                                "Saturday 21 March 2026  01:01:18 +0000 (0:00:00.060)       0:01:01.501 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  01:01:28 +0000 (0:00:09.418)       0:01:10.920 ******** ",
                                "=============================================================================== ",
                                "placement : Running placement bootstrap container ---------------------- 12.44s",
                                "placement : Restart placement-api container ----------------------------- 9.42s",
                                "service-ks-register : placement | Creating endpoints -------------------- 5.81s",
                                "service-ks-register : placement | Granting user roles ------------------- 4.24s",
                                "service-ks-register : placement | Creating projects --------------------- 3.76s",
                                "service-ks-register : placement | Creating users ------------------------ 3.72s",
                                "service-ks-register : placement | Creating services --------------------- 3.37s",
                                "service-ks-register : placement | Creating roles ------------------------ 3.05s",
                                "placement : Copying over placement.conf --------------------------------- 2.45s",
                                "placement : Creating placement databases user and setting permissions --- 2.14s",
                                "placement : Ensuring config directories exist --------------------------- 2.11s",
                                "service-cert-copy : placement | Copying over extra CA certificates ------ 1.99s",
                                "placement : Creating placement databases -------------------------------- 1.96s",
                                "placement : Copying over config.json files for services ----------------- 1.74s",
                                "placement : Copying over existing policy file --------------------------- 1.69s",
                                "placement : Copying over placement-api wsgi configuration --------------- 1.62s",
                                "placement : Copying over migrate-db.rc.j2 configuration ----------------- 1.50s",
                                "placement : Check placement containers ---------------------------------- 1.46s",
                                "service-cert-copy : placement | Copying over backend internal TLS key --- 1.29s",
                                "placement : include_tasks ----------------------------------------------- 1.09s",
                                "\u001b[32m2026-03-21 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 29d2a090-e26a-46c7-8734-92002c46d499 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 29d2a090-e26a-46c7-8734-92002c46d499 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 056bb715-91b6-4ef4-996f-99b605fab9e4 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  01:01:31 +0000 (0:00:00.167)       0:00:00.167 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  01:01:31 +0000 (0:00:00.296)       0:00:00.464 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_keystone_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_keystone_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_keystone_True)\u001b[0m",
                                "",
                                "PLAY [Wait for the Keystone service] *******************************************",
                                "",
                                "TASK [Waiting for Keystone public port to be UP] *******************************",
                                "Saturday 21 March 2026  01:01:31 +0000 (0:00:00.418)       0:00:00.882 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  01:01:32 +0000 (0:00:00.886)       0:00:01.768 ******** ",
                                "=============================================================================== ",
                                "Waiting for Keystone public port to be UP ------------------------------- 0.89s",
                                "Group hosts based on enabled services ----------------------------------- 0.42s",
                                "Group hosts based on Kolla action --------------------------------------- 0.30s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  00:58:43 +0000 (0:00:00.777)       0:00:00.778 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  00:58:44 +0000 (0:00:00.423)       0:00:01.202 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_designate_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_designate_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_designate_True)\u001b[0m",
                                "",
                                "PLAY [Apply role designate] ****************************************************",
                                "",
                                "TASK [designate : include_tasks] ***********************************************",
                                "Saturday 21 March 2026  00:58:44 +0000 (0:00:00.441)       0:00:01.643 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/designate/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating services] *********************",
                                "Saturday 21 March 2026  00:58:45 +0000 (0:00:00.959)       0:00:02.603 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate (dns))\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating endpoints] ********************",
                                "Saturday 21 March 2026  00:58:49 +0000 (0:00:04.012)       0:00:06.616 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> https://api-int.testbed.osism.xyz:9001 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> https://api.testbed.osism.xyz:9001 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating projects] *********************",
                                "Saturday 21 March 2026  00:58:56 +0000 (0:00:07.153)       0:00:13.769 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating users] ************************",
                                "Saturday 21 March 2026  00:59:00 +0000 (0:00:03.304)       0:00:17.073 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating roles] ************************",
                                "Saturday 21 March 2026  00:59:04 +0000 (0:00:04.272)       0:00:21.346 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Granting user roles] *******************",
                                "Saturday 21 March 2026  00:59:08 +0000 (0:00:03.648)       0:00:24.994 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> service -> admin)\u001b[0m",
                                "",
                                "TASK [designate : Ensuring config directories exist] ***************************",
                                "Saturday 21 March 2026  00:59:11 +0000 (0:00:03.672)       0:00:28.667 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : Check if policies shall be overwritten] **********************",
                                "Saturday 21 March 2026  00:59:14 +0000 (0:00:03.062)       0:00:31.729 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [designate : Set designate policy file] ***********************************",
                                "Saturday 21 March 2026  00:59:15 +0000 (0:00:00.116)       0:00:31.846 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [designate : include_tasks] ***********************************************",
                                "Saturday 21 March 2026  00:59:15 +0000 (0:00:00.243)       0:00:32.089 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/designate/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : designate | Copying over extra CA certificates] ******",
                                "Saturday 21 March 2026  00:59:15 +0000 (0:00:00.454)       0:00:32.544 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : designate | Copying over backend internal TLS certificate] ***",
                                "Saturday 21 March 2026  00:59:22 +0000 (0:00:06.700)       0:00:39.245 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : designate | Copying over backend internal TLS key] ***",
                                "Saturday 21 March 2026  00:59:23 +0000 (0:00:00.890)       0:00:40.135 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [designate : Copying over config.json files for services] *****************",
                                "Saturday 21 March 2026  00:59:24 +0000 (0:00:00.908)       0:00:41.044 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : Copying over designate.conf] *********************************",
                                "Saturday 21 March 2026  00:59:32 +0000 (0:00:08.447)       0:00:49.491 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : Copying over pools.yaml] *************************************",
                                "Saturday 21 March 2026  00:59:55 +0000 (0:00:22.578)       0:01:12.069 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m",
                                "",
                                "TASK [designate : Copying over named.conf] *************************************",
                                "Saturday 21 March 2026  01:00:02 +0000 (0:00:07.699)       0:01:19.769 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m",
                                "",
                                "TASK [designate : Copying over rndc.conf] **************************************",
                                "Saturday 21 March 2026  01:00:06 +0000 (0:00:03.243)       0:01:23.012 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : Copying over rndc.key] ***************************************",
                                "Saturday 21 March 2026  01:00:09 +0000 (0:00:03.611)       0:01:26.624 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : include_tasks] ***********************************************",
                                "Saturday 21 March 2026  01:00:13 +0000 (0:00:03.702)       0:01:30.327 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [designate : Copying over existing policy file] ***************************",
                                "Saturday 21 March 2026  01:00:14 +0000 (0:00:00.494)       0:01:30.821 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [designate : Check designate containers] **********************************",
                                "Saturday 21 March 2026  01:00:15 +0000 (0:00:01.300)       0:01:32.122 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : include_tasks] ***********************************************",
                                "Saturday 21 March 2026  01:00:19 +0000 (0:00:04.514)       0:01:36.637 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [designate : Creating Designate databases] ********************************",
                                "Saturday 21 March 2026  01:00:20 +0000 (0:00:00.383)       0:01:37.020 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate)\u001b[0m",
                                "",
                                "TASK [designate : Creating Designate databases user and setting permissions] ***",
                                "Saturday 21 March 2026  01:00:22 +0000 (0:00:01.977)       0:01:38.997 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0 -> {{ groups['designate-central'][0] }}]\u001b[0m",
                                "",
                                "TASK [designate : Running Designate bootstrap container] ***********************",
                                "Saturday 21 March 2026  01:00:24 +0000 (0:00:02.069)       0:01:41.067 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [designate : Flush handlers] **********************************************",
                                "Saturday 21 March 2026  01:00:37 +0000 (0:00:13.328)       0:01:54.395 ******** ",
                                "",
                                "TASK [designate : Flush handlers] **********************************************",
                                "Saturday 21 March 2026  01:00:37 +0000 (0:00:00.133)       0:01:54.529 ******** ",
                                "",
                                "TASK [designate : Flush handlers] **********************************************",
                                "Saturday 21 March 2026  01:00:37 +0000 (0:00:00.067)       0:01:54.597 ******** ",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-backend-bind9 container] ********",
                                "Saturday 21 March 2026  01:00:37 +0000 (0:00:00.063)       0:01:54.661 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-api container] ******************",
                                "Saturday 21 March 2026  01:00:47 +0000 (0:00:10.073)       0:02:04.734 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-central container] **************",
                                "Saturday 21 March 2026  01:00:59 +0000 (0:00:11.329)       0:02:16.063 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-producer container] *************",
                                "Saturday 21 March 2026  01:01:05 +0000 (0:00:05.900)       0:02:21.964 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-mdns container] *****************",
                                "Saturday 21 March 2026  01:01:13 +0000 (0:00:08.278)       0:02:30.243 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-worker container] ***************",
                                "Saturday 21 March 2026  01:01:23 +0000 (0:00:09.924)       0:02:40.167 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [designate : Non-destructive DNS pools update] ****************************",
                                "Saturday 21 March 2026  01:01:33 +0000 (0:00:10.134)       0:02:50.301 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=29  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  01:01:40 +0000 (0:00:06.906)       0:02:57.208 ******** ",
                                "=============================================================================== ",
                                "designate : Copying over designate.conf -------------------------------- 22.58s",
                                "designate : Running Designate bootstrap container ---------------------- 13.33s",
                                "designate : Restart designate-api container ---------------------------- 11.33s",
                                "designate : Restart designate-worker container ------------------------- 10.13s",
                                "designate : Restart designate-backend-bind9 container ------------------ 10.08s",
                                "designate : Restart designate-mdns container ---------------------------- 9.92s",
                                "designate : Copying over config.json files for services ----------------- 8.45s",
                                "designate : Restart designate-producer container ------------------------ 8.28s",
                                "designate : Copying over pools.yaml ------------------------------------- 7.70s",
                                "service-ks-register : designate | Creating endpoints -------------------- 7.15s",
                                "designate : Non-destructive DNS pools update ---------------------------- 6.91s",
                                "service-cert-copy : designate | Copying over extra CA certificates ------ 6.70s",
                                "designate : Restart designate-central container ------------------------- 5.90s",
                                "designate : Check designate containers ---------------------------------- 4.51s",
                                "service-ks-register : designate | Creating users ------------------------ 4.27s",
                                "service-ks-register : designate | Creating services --------------------- 4.01s",
                                "designate : Copying over rndc.key --------------------------------------- 3.71s",
                                "service-ks-register : designate | Granting user roles ------------------- 3.67s",
                                "service-ks-register : designate | Creating roles ------------------------ 3.65s",
                                "designate : Copying over rndc.conf -------------------------------------- 3.61s",
                                "\u001b[32m2026-03-21 01:01:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:01:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fda4417c-ac6b-473a-b4ba-ae594180dff9 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  01:01:43 +0000 (0:00:00.245)       0:00:00.245 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  01:01:44 +0000 (0:00:01.560)       0:00:01.805 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-rgw] *****************************************************",
                                "",
                                "TASK [ceph-rgw : include_tasks] ************************************************",
                                "Saturday 21 March 2026  01:01:46 +0000 (0:00:01.259)       0:00:03.065 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating services] **********************",
                                "Saturday 21 March 2026  01:01:48 +0000 (0:00:02.084)       0:00:05.150 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=swift (object-store))\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating endpoints] *********************",
                                "Saturday 21 March 2026  01:01:51 +0000 (0:00:03.696)       0:00:08.846 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=swift -> https://api-int.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=swift -> https://api.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating projects] **********************",
                                "Saturday 21 March 2026  01:01:58 +0000 (0:00:06.164)       0:00:15.011 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating users] *************************",
                                "Saturday 21 March 2026  01:02:01 +0000 (0:00:03.561)       0:00:18.573 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=ceph_rgw -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating roles] *************************",
                                "Saturday 21 March 2026  01:02:05 +0000 (0:00:04.273)       0:00:22.846 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=ResellerAdmin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Granting user roles] ********************",
                                "Saturday 21 March 2026  01:02:12 +0000 (0:00:06.495)       0:00:29.341 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=ceph_rgw -> service -> admin)\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  01:02:17 +0000 (0:00:05.043)       0:00:34.384 ******** ",
                                "=============================================================================== ",
                                "service-ks-register : ceph-rgw | Creating roles ------------------------- 6.49s",
                                "service-ks-register : ceph-rgw | Creating endpoints --------------------- 6.16s",
                                "service-ks-register : ceph-rgw | Granting user roles -------------------- 5.04s",
                                "service-ks-register : ceph-rgw | Creating users ------------------------- 4.27s",
                                "service-ks-register : ceph-rgw | Creating services ---------------------- 3.70s",
                                "service-ks-register : ceph-rgw | Creating projects ---------------------- 3.56s",
                                "ceph-rgw : include_tasks ------------------------------------------------ 2.08s",
                                "Group hosts based on Kolla action --------------------------------------- 1.56s",
                                "Group hosts based on enabled services ----------------------------------- 1.26s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  01:00:57 +0000 (0:00:00.570)       0:00:00.570 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  01:00:58 +0000 (0:00:00.636)       0:00:01.207 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_magnum_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_magnum_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_magnum_True)\u001b[0m",
                                "",
                                "PLAY [Apply role magnum] *******************************************************",
                                "",
                                "TASK [magnum : include_tasks] **************************************************",
                                "Saturday 21 March 2026  01:00:58 +0000 (0:00:00.377)       0:00:01.584 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/magnum/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating services] ************************",
                                "Saturday 21 March 2026  01:00:59 +0000 (0:00:00.962)       0:00:02.547 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum (container-infra))\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating endpoints] ***********************",
                                "Saturday 21 March 2026  01:01:03 +0000 (0:00:04.335)       0:00:06.882 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> https://api-int.testbed.osism.xyz:9511/v1 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> https://api.testbed.osism.xyz:9511/v1 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating projects] ************************",
                                "Saturday 21 March 2026  01:01:09 +0000 (0:00:06.159)       0:00:13.042 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating users] ***************************",
                                "Saturday 21 March 2026  01:01:13 +0000 (0:00:03.527)       0:00:16.570 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating roles] ***************************",
                                "Saturday 21 March 2026  01:01:16 +0000 (0:00:03.418)       0:00:19.988 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Granting user roles] **********************",
                                "Saturday 21 March 2026  01:01:19 +0000 (0:00:02.738)       0:00:22.727 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> service -> admin)\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum trustee domain] *********************************",
                                "Saturday 21 March 2026  01:01:23 +0000 (0:00:03.864)       0:00:26.591 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum trustee user] ***********************************",
                                "Saturday 21 March 2026  01:01:26 +0000 (0:00:03.160)       0:00:29.751 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum trustee user role] ******************************",
                                "Saturday 21 March 2026  01:01:30 +0000 (0:00:03.504)       0:00:33.256 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Ensuring config directories exist] ******************************",
                                "Saturday 21 March 2026  01:01:33 +0000 (0:00:03.333)       0:00:36.589 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [magnum : Check if policies shall be overwritten] *************************",
                                "Saturday 21 March 2026  01:01:35 +0000 (0:00:01.674)       0:00:38.264 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Set magnum policy file] *****************************************",
                                "Saturday 21 March 2026  01:01:35 +0000 (0:00:00.134)       0:00:38.398 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [magnum : Check if kubeconfig file is supplied] ***************************",
                                "Saturday 21 March 2026  01:01:35 +0000 (0:00:00.296)       0:00:38.695 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [magnum : Copying over kubeconfig file] ***********************************",
                                "Saturday 21 March 2026  01:01:36 +0000 (0:00:00.845)       0:00:39.540 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [magnum : Set magnum kubeconfig file's path] ******************************",
                                "Saturday 21 March 2026  01:01:38 +0000 (0:00:01.955)       0:00:41.496 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [magnum : include_tasks] **************************************************",
                                "Saturday 21 March 2026  01:01:38 +0000 (0:00:00.342)       0:00:41.838 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/magnum/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : magnum | Copying over extra CA certificates] *********",
                                "Saturday 21 March 2026  01:01:39 +0000 (0:00:00.926)       0:00:42.765 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : magnum | Copying over backend internal TLS certificate] ***",
                                "Saturday 21 March 2026  01:01:42 +0000 (0:00:02.445)       0:00:45.211 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : magnum | Copying over backend internal TLS key] ******",
                                "Saturday 21 March 2026  01:01:43 +0000 (0:00:01.078)       0:00:46.289 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}\u001b[32m2026-03-21 01:02:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask afe86918-e065-45e6-92ec-8b20bc35224d is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [magnum : Copying over config.json files for services] ********************",
                                "Saturday 21 March 2026  01:01:44 +0000 (0:00:01.347)       0:00:47.637 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [magnum : Copying over magnum.conf] ***************************************",
                                "Saturday 21 March 2026  01:01:47 +0000 (0:00:03.124)       0:00:50.761 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [magnum : Copying over existing policy file] ******************************",
                                "Saturday 21 March 2026  01:01:55 +0000 (0:00:07.861)       0:00:58.622 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [magnum : Check magnum containers] ****************************************",
                                "Saturday 21 March 2026  01:01:56 +0000 (0:00:01.153)       0:00:59.776 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [magnum : include_tasks] **************************************************",
                                "Saturday 21 March 2026  01:01:58 +0000 (0:00:02.000)       0:01:01.776 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum database] ***************************************",
                                "Saturday 21 March 2026  01:01:58 +0000 (0:00:00.306)       0:01:02.083 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum database user and setting permissions] **********",
                                "Saturday 21 March 2026  01:02:01 +0000 (0:00:02.386)       0:01:04.469 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Running Magnum bootstrap container] *****************************",
                                "Saturday 21 March 2026  01:02:04 +0000 (0:00:02.693)       0:01:07.162 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Flush handlers] *************************************************",
                                "Saturday 21 March 2026  01:02:21 +0000 (0:00:17.509)       0:01:24.671 ******** ",
                                "",
                                "TASK [magnum : Flush handlers] *************************************************",
                                "Saturday 21 March 2026  01:02:21 +0000 (0:00:00.201)       0:01:24.873 ******** ",
                                "",
                                "TASK [magnum : Flush handlers] *************************************************",
                                "Saturday 21 March 2026  01:02:21 +0000 (0:00:00.060)       0:01:24.933 ******** ",
                                "",
                                "RUNNING HANDLER [magnum : Restart magnum-api container] ************************",
                                "Saturday 21 March 2026  01:02:21 +0000 (0:00:00.076)       0:01:25.010 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [magnum : Restart magnum-conductor container] ******************",
                                "Saturday 21 March 2026  01:02:37 +0000 (0:00:15.419)       0:01:40.430 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  01:02:45 +0000 (0:00:07.918)       0:01:48.348 ******** ",
                                "=============================================================================== ",
                                "magnum : Running Magnum bootstrap container ---------------------------- 17.51s",
                                "magnum : Restart magnum-api container ---------------------------------- 15.42s",
                                "magnum : Restart magnum-conductor container ----------------------------- 7.92s",
                                "magnum : Copying over magnum.conf --------------------------------------- 7.86s",
                                "service-ks-register : magnum | Creating endpoints ----------------------- 6.16s",
                                "service-ks-register : magnum | Creating services ------------------------ 4.34s",
                                "service-ks-register : magnum | Granting user roles ---------------------- 3.86s",
                                "service-ks-register : magnum | Creating projects ------------------------ 3.53s",
                                "magnum : Creating Magnum trustee user ----------------------------------- 3.50s",
                                "service-ks-register : magnum | Creating users --------------------------- 3.42s",
                                "magnum : Creating Magnum trustee user role ------------------------------ 3.33s",
                                "magnum : Creating Magnum trustee domain --------------------------------- 3.16s",
                                "magnum : Copying over config.json files for services -------------------- 3.12s",
                                "service-ks-register : magnum | Creating roles --------------------------- 2.74s",
                                "magnum : Creating Magnum database user and setting permissions ---------- 2.69s",
                                "service-cert-copy : magnum | Copying over extra CA certificates --------- 2.45s",
                                "magnum : Creating Magnum database --------------------------------------- 2.39s",
                                "magnum : Check magnum containers ---------------------------------------- 2.00s",
                                "magnum : Copying over kubeconfig file ----------------------------------- 1.95s",
                                "magnum : Ensuring config directories exist ------------------------------ 1.67s",
                                "\u001b[32m2026-03-21 01:02:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30517df1-a1e5-4332-bb1e-8fe94c810db9 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  00:58:44 +0000 (0:00:00.394)       0:00:00.394 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  00:58:45 +0000 (0:00:00.951)       0:00:01.346 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_neutron_True)\u001b[0m",
                                "",
                                "PLAY [Apply role neutron] ******************************************************",
                                "",
                                "TASK [neutron : include_tasks] *************************************************",
                                "Saturday 21 March 2026  00:58:46 +0000 (0:00:00.909)       0:00:02.255 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/neutron/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [neutron : Get container facts] *******************************************",
                                "Saturday 21 March 2026  00:58:47 +0000 (0:00:01.182)       0:00:03.438 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Get container volume facts] ************************************",
                                "Saturday 21 March 2026  00:58:48 +0000 (0:00:01.362)       0:00:04.800 ******** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Check for ML2/OVN presence] ************************************",
                                "Saturday 21 March 2026  00:58:50 +0000 (0:00:01.127)       0:00:05.927 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [neutron : Check for ML2/OVS presence] ************************************",
                                "Saturday 21 March 2026  00:58:50 +0000 (0:00:00.579)       0:00:06.507 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating services] ***********************",
                                "Saturday 21 March 2026  00:58:51 +0000 (0:00:00.613)       0:00:07.121 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron (network))\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating endpoints] **********************",
                                "Saturday 21 March 2026  00:58:55 +0000 (0:00:03.942)       0:00:11.063 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> https://api-int.testbed.osism.xyz:9696 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> https://api.testbed.osism.xyz:9696 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating projects] ***********************",
                                "Saturday 21 March 2026  00:59:01 +0000 (0:00:06.006)       0:00:17.069 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating users] **************************",
                                "Saturday 21 March 2026  00:59:04 +0000 (0:00:03.309)       0:00:20.379 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating roles] **************************",
                                "Saturday 21 March 2026  00:59:08 +0000 (0:00:04.289)       0:00:24.669 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Granting user roles] *********************",
                                "Saturday 21 March 2026  00:59:11 +0000 (0:00:03.228)       0:00:27.897 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service -> admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service -> service)\u001b[0m",
                                "",
                                "TASK [neutron : include_tasks] *************************************************",
                                "Saturday 21 March 2026  00:59:19 +0000 (0:00:07.300)       0:00:35.198 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Load and persist kernel modules] *****************************************",
                                "Saturday 21 March 2026  00:59:19 +0000 (0:00:00.463)       0:00:35.661 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Check IPv6 support] ********************************************",
                                "Saturday 21 March 2026  00:59:22 +0000 (0:00:02.679)       0:00:38.340 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Setting sysctl values] ***************************************************",
                                "Saturday 21 March 2026  00:59:23 +0000 (0:00:01.464)       0:00:39.805 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Ensuring config directories exist] *****************************",
                                "Saturday 21 March 2026  00:59:27 +0000 (0:00:03.197)       0:00:43.002 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [neutron : Check if extra ml2 plugins exists] *****************************",
                                "Saturday 21 March 2026  00:59:29 +0000 (0:00:02.811)       0:00:45.814 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/neutron/plugins/' path\u001b[0m",
                                "\u001b[1;35mdue to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/neutron/plugins/' is not\u001b[0m",
                                "\u001b[1;35ma directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [neutron : include_tasks] *************************************************",
                                "Saturday 21 March 2026  00:59:30 +0000 (0:00:00.869)       0:00:46.684 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/neutron/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [service-cert-copy : neutron | Copying over extra CA certificates] ********",
                                "Saturday 21 March 2026  00:59:32 +0000 (0:00:01.278)       0:00:47.962 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : neutron | Copying over backend internal TLS certificate] ***",
                                "Saturday 21 March 2026  00:59:37 +0000 (0:00:05.311)       0:00:53.274 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : neutron | Copying over backend internal TLS key] *****",
                                "Saturday 21 March 2026  00:59:40 +0000 (0:00:03.103)       0:00:56.378 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Creating TLS backend PEM File] *********************************",
                                "Saturday 21 March 2026  00:59:43 +0000 (0:00:03.223)       0:00:59.602 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Check if policies shall be overwritten] ************************",
                                "Saturday 21 March 2026  00:59:46 +0000 (0:00:02.527)       0:01:02.129 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [neutron : Set neutron policy file] ***************************************",
                                "Saturday 21 March 2026  00:59:46 +0000 (0:00:00.222)       0:01:02.352 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over existing policy file] *****************************",
                                "Saturday 21 March 2026  00:59:46 +0000 (0:00:00.553)       0:01:02.906 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over config.json files for services] *******************",
                                "Saturday 21 March 2026  00:59:50 +0000 (0:00:03.954)       0:01:06.860 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron.conf] *************************************",
                                "Saturday 21 March 2026  00:59:55 +0000 (0:00:04.097)       0:01:10.958 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron_vpnaas.conf] ******************************",
                                "Saturday 21 March 2026  01:00:03 +0000 (0:00:08.275)       0:01:19.234 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over ssh key] ******************************************",
                                "Saturday 21 March 2026  01:00:05 +0000 (0:00:02.502)       0:01:21.736 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over ml2_conf.ini] *************************************",
                                "Saturday 21 March 2026  01:00:09 +0000 (0:00:03.230)       0:01:24.967 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "",
                                "TASK [neutron : Copying over linuxbridge_agent.ini] ****************************",
                                "Saturday 21 March 2026  01:00:13 +0000 (0:00:03.956)       0:01:28.923 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over openvswitch_agent.ini] ****************************",
                                "Saturday 21 March 2026  01:00:15 +0000 (0:00:02.400)       0:01:31.324 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over sriov_agent.ini] **********************************",
                                "Saturday 21 March 2026  01:00:17 +0000 (0:00:02.514)       0:01:33.839 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over mlnx_agent.ini] ***********************************",
                                "Saturday 21 March 2026  01:00:19 +0000 (0:00:01.951)       0:01:35.790 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over eswitchd.conf] ************************************",
                                "Saturday 21 March 2026  01:00:21 +0000 (0:00:02.063)       0:01:37.854 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over dhcp_agent.ini] ***********************************",
                                "Saturday 21 March 2026  01:00:23 +0000 (0:00:02.029)       0:01:39.884 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over dnsmasq.conf] *************************************",
                                "Saturday 21 March 2026  01:00:25 +0000 (0:00:01.873)       0:01:41.757 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over l3_agent.ini] *************************************",
                                "Saturday 21 March 2026  01:00:27 +0000 (0:00:01.701)       0:01:43.458 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over fwaas_driver.ini] *********************************",
                                "Saturday 21 March 2026  01:00:29 +0000 (0:00:01.914)       0:01:45.373 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over metadata_agent.ini] *******************************",
                                "Saturday 21 March 2026  01:00:31 +0000 (0:00:01.736)       0:01:47.110 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron_ovn_metadata_agent.ini] *******************",
                                "Saturday 21 March 2026  01:00:33 +0000 (0:00:02.401)       0:01:49.511 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over metering_agent.ini] *******************************",
                                "Saturday 21 March 2026  01:00:36 +0000 (0:00:03.075)       0:01:52.587 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over ironic_neutron_agent.ini] *************************",
                                "Saturday 21 March 2026  01:00:38 +0000 (0:00:02.257)       0:01:54.845 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over bgp_dragent.ini] **********************************",
                                "Saturday 21 March 2026  01:00:41 +0000 (0:00:02.655)       0:01:57.500 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over ovn_agent.ini] ************************************",
                                "Saturday 21 March 2026  01:00:44 +0000 (0:00:03.117)       0:02:00.618 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over nsx.ini] ******************************************",
                                "Saturday 21 March 2026  01:00:46 +0000 (0:00:01.630)       0:02:02.248 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copy neutron-l3-agent-wrapper script] **************************",
                                "Saturday 21 March 2026  01:00:49 +0000 (0:00:02.786)       0:02:05.035 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over extra ml2 plugins] ********************************",
                                "Saturday 21 March 2026  01:00:52 +0000 (0:00:02.875)       0:02:07.911 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron-tls-proxy.cfg] ****************************",
                                "Saturday 21 March 2026  01:00:54 +0000 (0:00:02.574)       0:02:10.485 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron_taas.conf] ********************************",
                                "Saturday 21 March 2026  01:00:56 +0000 (0:00:01.909)       0:02:12.395 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Check neutron containers] **************************************",
                                "Saturday 21 March 2026  01:00:58 +0000 (0:00:02.257)       0:02:14.652 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [neutron : include_tasks] *************************************************",
                                "Saturday 21 March 2026  01:01:02 +0000 (0:00:03.413)       0:02:18.066 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Creating Neutron database] *************************************",
                                "Saturday 21 March 2026  01:01:02 +0000 (0:00:00.561)       0:02:18.627 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [neutron : Creating Neutron database user and setting permissions] ********",
                                "Saturday 21 March 2026  01:01:04 +0000 (0:00:01.941)       0:02:20.568 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [neutron : Running Neutron bootstrap container] ***************************",
                                "Saturday 21 March 2026  01:01:06 +0000 (0:00:02.320)       0:02:22.888 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Saturday 21 March 2026  01:01:43 +0000 (0:00:37.007)       0:02:59.896 ******** ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Saturday 21 March 2026  01:01:44 +0000 (0:00:00.066)       0:02:59.963 ******** ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Saturday 21 March 2026  01:01:44 +0000 (0:00:00.112)       0:03:00.075 ******** ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Saturday 21 March 2026  01:01:44 +0000 (0:00:00.118)       0:03:00.193 ******** ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Saturday 21 March 2026  01:01:44 +0000 (0:00:00.112)       0:03:00.306 ******** ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Saturday 21 March 2026  01:01:44 +0000 (0:00:00.070)       0:03:00.377 ******** ",
                                "",
                                "RUNNING HANDLER [neutron : Restart neutron-server container] *******************",
                                "Saturday 21 March 2026  01:01:44 +0000 (0:00:00.066)       0:03:00.443 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [neutron : Restart neutron-ovn-metadata-agent container] *******",
                                "Saturday 21 March 2026  01:02:11 +0000 (0:00:26.633)       0:03:27.077 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=31  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=31  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  01:02:53 +0000 (0:00:42.713)       0:04:09.790 ******** ",
                                "=============================================================================== ",
                                "neutron : Restart neutron-ovn-metadata-agent container ----------------- 42.71s",
                                "neutron : Running Neutron bootstrap container -------------------------- 37.01s",
                                "neutron : Restart neutron-server container ----------------------------- 26.63s",
                                "neutron : Copying over neutron.conf ------------------------------------- 8.28s",
                                "service-ks-register : neutron | Granting user roles --------------------- 7.30s",
                                "service-ks-register : neutron | Creating endpoints ---------------------- 6.01s",
                                "service-cert-copy : neutron | Copying over extra CA certificates -------- 5.31s",
                                "service-ks-register : neutron | Creating users -------------------------- 4.29s",
                                "neutron : Copying over config.json files for services ------------------- 4.10s",
                                "neutron : Copying over ml2_conf.ini ------------------------------------- 3.96s",
                                "neutron : Copying over existing policy file ----------------------------- 3.95s",
                                "service-ks-register : neutron | Creating services ----------------------- 3.94s",
                                "neutron : Check neutron containers -------------------------------------- 3.41s",
                                "service-ks-register : neutron | Creating projects ----------------------- 3.31s",
                                "neutron : Copying over ssh key ------------------------------------------ 3.23s",
                                "service-ks-register : neutron | Creating roles -------------------------- 3.23s",
                                "service-cert-copy : neutron | Copying over backend internal TLS key ----- 3.22s",
                                "Setting sysctl values --------------------------------------------------- 3.20s",
                                "neutron : Copying over bgp_dragent.ini ---------------------------------- 3.12s",
                                "service-cert-copy : neutron | Copying over backend internal TLS certificate --- 3.10s",
                                "\u001b[32m2026-03-21 01:02:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:02:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9efad5cd-fb73-405a-960f-ee4602a8ebe0 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  01:01:35 +0000 (0:00:00.288)       0:00:00.288 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  01:01:36 +0000 (0:00:00.631)       0:00:00.920 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_prometheus_True)\u001b[0m",
                                "",
                                "PLAY [Apply role prometheus] ***************************************************",
                                "",
                                "TASK [prometheus : include_tasks] **********************************************",
                                "Saturday 21 March 2026  01:01:37 +0000 (0:00:00.765)       0:00:01.685 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/prometheus/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [prometheus : Ensuring config directories exist] **************************",
                                "Saturday 21 March 2026  01:01:38 +0000 (0:00:01.115)       0:00:02.801 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [prometheus : include_tasks] **********************************************",
                                "Saturday 21 March 2026  01:01:41 +0000 (0:00:03.266)       0:00:06.068 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/prometheus/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [service-cert-copy : prometheus | Copying over extra CA certificates] *****",
                                "Saturday 21 March 2026  01:01:43 +0000 (0:00:01.310)       0:00:07.378 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : prometheus | Copying over backend internal TLS certificate] ***",
                                "Saturday 21 March 2026  01:01:49 +0000 (0:00:06.480)       0:00:13.859 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : prometheus | Copying over backend internal TLS key] ***",
                                "Saturday 21 March 2026  01:01:51 +0000 (0:00:02.313)       0:00:16.172 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over config.json files] *****************************",
                                "Saturday 21 March 2026  01:01:54 +0000 (0:00:02.707)       0:00:18.880 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [prometheus : Find custom prometheus alert rules files] *******************",
                                "Saturday 21 March 2026  01:02:00 +0000 (0:00:06.364)       0:00:25.244 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over custom prometheus alert rules files] ***********",
                                "Saturday 21 March 2026  01:02:01 +0000 (0:00:01.017)       0:00:26.262 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1106499, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0519824, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1106499, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0519824, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1106499, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0519824, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1106527, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.061685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1106499, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0519824, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1106527, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.061685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1106499, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0519824, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1106527, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.061685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1106493, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.050685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1106493, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.050685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1106499, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0519824, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1106527, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.061685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1106499, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0519824, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1106527, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.061685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1106493, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.050685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1106512, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.055685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1106527, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.061685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1106493, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.050685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1106512, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.055685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1106493, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.050685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1106489, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04872, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1106512, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.055685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1106493, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.050685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1106501, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0524561, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1106512, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.055685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1106527, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.061685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1106489, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04872, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1106512, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.055685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1106489, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04872, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1106512, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.055685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1106510, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0552485, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1106489, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04872, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1106501, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0524561, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1106489, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04872, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1106501, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0524561, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1106503, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.052685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1106489, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04872, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1106501, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0524561, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1106501, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0524561, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1106510, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0552485, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1106493, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.050685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1106510, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0552485, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1106510, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0552485, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1106503, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.052685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1106503, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.052685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1106510, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0552485, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1106501, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0524561, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1106498, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0515304, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1106498, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0515304, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1106503, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.052685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1106498, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0515304, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1106510, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0552485, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1106503, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.052685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106526, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0597665, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1106498, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0515304, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106526, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0597665, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106526, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0597665, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1106498, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0515304, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1106503, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.052685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106526, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0597665, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106486, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0478735, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106526, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0597665, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106486, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0478735, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106486, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0478735, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106486, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0478735, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1106512, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.055685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106486, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0478735, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1106549, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0655851, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1106498, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0515304, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1106549, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0655851, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1106549, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0655851, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1106524, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.058685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1106549, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0655851, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1106549, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0655851, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1106524, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.058685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106526, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0597665, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106490, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0489995, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1106524, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.058685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1106524, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.058685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106486, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0478735, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106490, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0489995, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1106489, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04872, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1106524, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.058685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1106488, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04834, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106490, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0489995, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1106549, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0655851, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1106507, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0546963, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106490, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0489995, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1106488, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04834, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1106488, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04834, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1106505, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0536127, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106490, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0489995, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1106488, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04834, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1106524, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.058685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1106507, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0546963, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1106507, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0546963, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106490, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0489995, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1106545, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0648096, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1106488, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04834, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1106501, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0524561, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1106507, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0546963, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1106505, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0536127, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1106505, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0536127, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1106505, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0536127, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1106488, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04834, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1106507, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0546963, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1106545, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0648096, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1106545, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0648096, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1106545, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0648096, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1106507, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0546963, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1106505, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0536127, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1106510, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0552485, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1106545, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0648096, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1106505, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0536127, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1106545, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0648096, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1106503, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.052685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1106498, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0515304, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106526, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0597665, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106486, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0478735, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1106549, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0655851, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1106524, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.058685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1106490, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0489995, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1106488, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.04834, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1106507, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0546963, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1106505, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0536127, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1106545, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0648096, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "",
                                "TASK [prometheus : Find prometheus common config overrides] ********************",
                                "Saturday 21 March 2026  01:02:27 +0000 (0:00:25.372)       0:00:51.634 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Find prometheus host config overrides] **********************",
                                "Saturday 21 March 2026  01:02:28 +0000 (0:00:00.732)       0:00:52.367 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mmanager/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mmanager/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-0/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-0/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-1/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-1/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-3/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-3/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-2/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-2/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-4/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-4/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-5/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-5/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over prometheus config file] ************************",
                                "Saturday 21 March 2026  01:02:29 +0000 (0:00:01.525)       0:00:53.892 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2)\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over prometheus web config file] ********************",
                                "Saturday 21 March 2026  01:02:43 +0000 (0:00:13.824)       0:01:07.717 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2)\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over prometheus alertmanager config file] ***********",
                                "Saturday 21 March 2026  01:02:47 +0000 (0:00:03.647)       0:01:11.365 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Find custom Alertmanager alert notification templates] ******",
                                "Saturday 21 March 2026  01:02:48 +0000 (0:00:01.640)       0:01:13.005 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over custom Alertmanager alert notification templates] ***",
                                "Saturday 21 March 2026  01:02:49 +0000 (0:00:00.765)       0:01:13.771 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over my.cnf for mysqld_exporter] ********************",
                                "Saturday 21 March 2026  01:02:50 +0000 (0:00:00.869)       0:01:14.640 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying cloud config file for openstack exporter] ***********",
                                "Saturday 21 March 2026  01:02:52 +0000 (0:00:02.642)       0:01:17.283 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying config file for blackbox exporter] ******************",
                                "Saturday 21 March 2026  01:02:54 +0000 (0:00:01.875)       0:01:19.159 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2)\u001b[0m",
                                "",
                                "TASK [prometheus : Find extra prometheus server config files] ******************",
                                "Saturday 21 March 2026  01:02:56 +0000 (0:00:01.585)       0:01:20.745 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/extras/' path\u001b[0m",
                                "\u001b[1;35mdue to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/extras/' is\u001b[0m",
                                "\u001b[1;35mnot a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Create subdirectories for extra config files] ***************",
                                "Saturday 21 March 2026  01:02:57 +0000 (0:00:01.123)       0:01:21.868 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Template extra prometheus server config files] **************",
                                "Saturday 21 March 2026  01:02:58 +0000 (0:00:00.656)       0:01:22.525 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Check prometheus containers] ********************************",
                                "Saturday 21 March 2026  01:02:59 +0000 (0:00:00.976)       0:01:23.502 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [prometheus : Creating prometheus database user and setting permissions] ***",
                                "Saturday 21 March 2026  01:03:03 +0000 (0:00:04.747)       0:01:28.249 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Saturday 21 March 2026  01:03:04 +0000 (0:00:00.786)       0:01:29.036 ******** ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Saturday 21 March 2026  01:03:04 +0000 (0:00:00.052)       0:01:29.088 ******** ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Saturday 21 March 2026  01:03:04 +0000 (0:00:00.049)       0:01:29.137 ******** ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Saturday 21 March 2026  01:03:04 +0000 (0:00:00.050)       0:01:29.188 ******** ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Saturday 21 March 2026  01:03:04 +0000 (0:00:00.050)       0:01:29.238 ******** ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Saturday 21 March 2026  01:03:04 +0000 (0:00:00.050)       0:01:29.289 ******** ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Saturday 21 March 2026  01:03:05 +0000 (0:00:00.051)       0:01:29.340 ******** ",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-server container] *************",
                                "Saturday 21 March 2026  01:03:05 +0000 (0:00:00.069)       0:01:29.410 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-node-exporter container] ******",
                                "Saturday 21 March 2026  01:03:20 +0000 (0:00:15.705)       0:01:45.115 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-mysqld-exporter container] ****",
                                "Saturday 21 March 2026  01:03:33 +0000 (0:00:12.888)       0:01:58.004 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-memcached-exporter container] ***",
                                "Saturday 21 March 2026  01:03:43 +0000 (0:00:09.801)       0:02:07.806 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-cadvisor container] ***********",
                                "Saturday 21 March 2026  01:03:49 +0000 (0:00:05.972)       0:02:13.778 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-alertmanager container] *******",
                                "Saturday 21 March 2026  01:04:04 +0000 (0:00:15.348)       0:02:29.126 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-elasticsearch-exporter container] ***",
                                "Saturday 21 March 2026  01:04:11 +0000 (0:00:06.529)       0:02:35.656 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-blackbox-exporter container] ***",
                                "Saturday 21 March 2026  01:04:20 +0000 (0:00:09.515)       0:02:45.171 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-libvirt-exporter container] ***",
                                "Saturday 21 March 2026  01:04:30 +0000 (0:00:09.884)       0:02:55.056 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  01:04:39 +0000 (0:00:08.892)       0:03:03.948 ******** ",
                                "=============================================================================== ",
                                "prometheus : Copying over custom prometheus alert rules files ---------- 25.37s",
                                "prometheus : Restart prometheus-server container ----------------------- 15.71s",
                                "prometheus : Restart prometheus-cadvisor container --------------------- 15.35s",
                                "prometheus : Copying over prometheus config file ----------------------- 13.82s",
                                "prometheus : Restart prometheus-node-exporter container ---------------- 12.89s",
                                "prometheus : Restart prometheus-blackbox-exporter container ------------- 9.88s",
                                "prometheus : Restart prometheus-mysqld-exporter container --------------- 9.80s",
                                "prometheus : Restart prometheus-elasticsearch-exporter container -------- 9.52s",
                                "prometheus : Restart prometheus-libvirt-exporter container -------------- 8.89s",
                                "prometheus : Restart prometheus-alertmanager container ------------------ 6.53s",
                                "service-cert-copy : prometheus | Copying over extra CA certificates ----- 6.48s",
                                "prometheus : Copying over config.json files ----------------------------- 6.36s",
                                "prometheus : Restart prometheus-memcached-exporter container ------------ 5.97s",
                                "prometheus : Check prometheus containers -------------------------------- 4.75s",
                                "prometheus : Copying over prometheus web config file -------------------- 3.65s",
                                "prometheus : Ensuring config directories exist -------------------------- 3.27s",
                                "service-cert-copy : prometheus | Copying over backend internal TLS key --- 2.71s",
                                "prometheus : Copying over my.cnf for mysqld_exporter -------------------- 2.64s",
                                "service-cert-copy : prometheus | Copying over backend internal TLS certificate --- 2.31s",
                                "prometheus : Copying cloud config file for openstack exporter ----------- 1.88s",
                                "\u001b[32m2026-03-21 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:04:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 963eff64-84a8-4ab6-8e40-0acaaa37ff53 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  01:02:20 +0000 (0:00:00.312)       0:00:00.312 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  01:02:21 +0000 (0:00:00.265)       0:00:00.578 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_glance_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_glance_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_glance_True)\u001b[0m",
                                "",
                                "PLAY [Apply role glance] *******************************************************",
                                "",
                                "TASK [glance : include_tasks] **************************************************",
                                "Saturday 21 March 2026  01:02:21 +0000 (0:00:00.258)       0:00:00.836 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/glance/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating services] ************************",
                                "Saturday 21 March 2026  01:02:22 +0000 (0:00:00.915)       0:00:01.751 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance (image))\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating endpoints] ***********************",
                                "Saturday 21 March 2026  01:02:25 +0000 (0:00:03.327)       0:00:05.079 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> https://api-int.testbed.osism.xyz:9292 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> https://api.testbed.osism.xyz:9292 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating projects] ************************",
                                "Saturday 21 March 2026  01:02:31 +0000 (0:00:05.636)       0:00:10.715 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating users] ***************************",
                                "Saturday 21 March 2026  01:02:33 +0000 (0:00:02.671)       0:00:13.387 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating roles] ***************************",
                                "Saturday 21 March 2026  01:02:37 +0000 (0:00:03.388)       0:00:16.775 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Granting user roles] **********************",
                                "Saturday 21 March 2026  01:02:40 +0000 (0:00:03.066)       0:00:19.841 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> service -> admin)\u001b[0m",
                                "",
                                "TASK [glance : Ensuring config directories exist] ******************************",
                                "Saturday 21 March 2026  01:02:44 +0000 (0:00:03.645)       0:00:23.487 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "",
                                "TASK [glance : include_tasks] **************************************************",
                                "Saturday 21 March 2026  01:02:48 +0000 (0:00:04.504)       0:00:27.994 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/glance/tasks/external_ceph.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [glance : Ensuring glance service ceph config subdir exists] **************",
                                "Saturday 21 March 2026  01:02:49 +0000 (0:00:00.580)       0:00:28.574 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Copy over multiple ceph configs for Glance] *********************",
                                "Saturday 21 March 2026  01:02:53 +0000 (0:00:04.261)       0:00:32.836 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "",
                                "TASK [glance : Copy over ceph Glance keyrings] *********************************",
                                "Saturday 21 March 2026  01:02:55 +0000 (0:00:02.022)       0:00:34.858 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "",
                                "TASK [glance : Ensuring config directory has correct owner and permission] *****",
                                "Saturday 21 March 2026  01:02:56 +0000 (0:00:01.542)       0:00:36.401 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Check if policies shall be overwritten] *************************",
                                "Saturday 21 March 2026  01:02:57 +0000 (0:00:00.669)       0:00:37.071 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Set glance policy file] *****************************************",
                                "Saturday 21 March 2026  01:02:57 +0000 (0:00:00.097)       0:00:37.168 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : include_tasks] **************************************************",
                                "Saturday 21 March 2026  01:02:58 +0000 (0:00:00.313)       0:00:37.482 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/glance/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : glance | Copying over extra CA certificates] *********",
                                "Saturday 21 March 2026  01:02:58 +0000 (0:00:00.586)       0:00:38.069 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : glance | Copying over backend internal TLS certificate] ***",
                                "Saturday 21 March 2026  01:03:03 +0000 (0:00:04.907)       0:00:42.977 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : glance | Copying over backend internal TLS key] ******",
                                "Saturday 21 March 2026  01:03:06 +0000 (0:00:02.683)       0:00:45.660 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Creating TLS backend PEM File] **********************************",
                                "Saturday 21 March 2026  01:03:09 +0000 (0:00:03.125)       0:00:48.786 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Copying over config.json files for services] ********************",
                                "Saturday 21 March 2026  01:03:12 +0000 (0:00:03.485)       0:00:52.272 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "",
                                "TASK [glance : Copying over glance-api.conf] ***********************************",
                                "Saturday 21 March 2026  01:03:17 +0000 (0:00:04.225)       0:00:56.497 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [glance : Copying over glance-cache.conf for glance_api] ******************",
                                "Saturday 21 March 2026  01:03:25 +0000 (0:00:08.101)       0:01:04.599 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Copying over glance-swift.conf for glance_api] ******************",
                                "Saturday 21 March 2026  01:03:29 +0000 (0:00:04.519)       0:01:09.118 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Copying over glance-image-import.conf] **************************",
                                "Saturday 21 March 2026  01:03:33 +0000 (0:00:03.360)       0:01:12.479 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Copying over property-protections-rules.conf] *******************",
                                "Saturday 21 March 2026  01:03:36 +0000 (0:00:03.940)       0:01:16.420 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Copying over existing policy file] ******************************",
                                "Saturday 21 March 2026  01:03:39 +0000 (0:00:02.907)       0:01:19.328 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Copying over glance-haproxy-tls.cfg] ****************************",
                                "Saturday 21 March 2026  01:03:40 +0000 (0:00:00.371)       0:01:19.699 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Generating 'hostnqn' file for glance_api] ***********************",
                                "Saturday 21 March 2026  01:03:43 +0000 (0:00:03.511)       0:01:23.211 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Generating 'hostid' file for glance_api] ************************",
                                "Saturday 21 March 2026  01:03:48 +0000 (0:00:04.657)       0:01:27.868 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Check glance containers] ****************************************",
                                "Saturday 21 March 2026  01:03:54 +0000 (0:00:06.562)       0:01:34.430 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "",
                                "TASK [glance : include_tasks] **************************************************",
                                "Saturday 21 March 2026  01:04:00 +0000 (0:00:05.160)       0:01:39.591 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Creating Glance database] ***************************************",
                                "Saturday 21 March 2026  01:04:00 +0000 (0:00:00.443)       0:01:40.034 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Creating Glance database user and setting permissions] **********",
                                "Saturday 21 March 2026  01:04:02 +0000 (0:00:01.860)       0:01:41.894 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Enable log_bin_trust_function_creators function] ****************",
                                "Saturday 21 March 2026  01:04:04 +0000 (0:00:01.834)       0:01:43.729 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Running Glance bootstrap container] *****************************",
                                "Saturday 21 March 2026  01:04:05 +0000 (0:00:01.731)       0:01:45.460 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Disable log_bin_trust_function_creators function] ***************",
                                "Saturday 21 March 2026  01:04:32 +0000 (0:00:26.793)       0:02:12.254 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Flush handlers] *************************************************",
                                "Saturday 21 March 2026  01:04:34 +0000 (0:00:01.715)       0:02:13.970 ******** ",
                                "",
                                "TASK [glance : Flush handlers] *************************************************",
                                "Saturday 21 March 2026  01:04:34 +0000 (0:00:00.064)       0:02:14.034 ******** ",
                                "",
                                "TASK [glance : Flush handlers] *************************************************",
                                "Saturday 21 March 2026  01:04:34 +0000 (0:00:00.061)       0:02:14.096 ******** ",
                                "",
                                "RUNNING HANDLER [glance : Restart glance-api container] ************************",
                                "Saturday 21 March 2026  01:04:34 +0000 (0:00:00.065)       0:02:14.162 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=14  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=13  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=13  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  01:05:01 +0000 (0:00:26.327)       0:02:40.490 ******** ",
                                "=============================================================================== ",
                                "glance : Running Glance bootstrap container ---------------------------- 26.79s",
                                "glance : Restart glance-api container ---------------------------------- 26.33s",
                                "glance : Copying over glance-api.conf ----------------------------------- 8.10s",
                                "glance : Generating 'hostid' file for glance_api ------------------------ 6.56s",
                                "service-ks-register : glance | Creating endpoints ----------------------- 5.64s",
                                "glance : Check glance containers ---------------------------------------- 5.16s",
                                "service-cert-copy : glance | Copying over extra CA certificates --------- 4.91s",
                                "glance : Generating 'hostnqn' file for glance_api ----------------------- 4.66s",
                                "glance : Copying over glance-cache.conf for glance_api ------------------ 4.52s",
                                "glance : Ensuring config directories exist ------------------------------ 4.50s",
                                "glance : Ensuring glance service ceph config subdir exists -------------- 4.26s",
                                "glance : Copying over config.json files for services -------------------- 4.23s",
                                "glance : Copying over glance-image-import.conf -------------------------- 3.94s",
                                "service-ks-register : glance | Granting user roles ---------------------- 3.65s",
                                "glance : Copying over glance-haproxy-tls.cfg ---------------------------- 3.51s",
                                "glance : Creating TLS backend PEM File ---------------------------------- 3.49s",
                                "service-ks-register : glance | Creating users --------------------------- 3.39s",
                                "glance : Copying over glance-swift.conf for glance_api ------------------ 3.36s",
                                "service-ks-register : glance | Creating services ------------------------ 3.33s",
                                "service-cert-copy : glance | Copying over backend internal TLS key ------ 3.13s",
                                "\u001b[32m2026-03-21 01:05:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11dbe5b3-dffb-4819-be0d-e8ada268beba is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  01:02:49 +0000 (0:00:00.269)       0:00:00.269 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  01:02:50 +0000 (0:00:00.392)       0:00:00.661 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_cinder_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_cinder_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_cinder_True)\u001b[0m",
                                "",
                                "PLAY [Apply role cinder] *******************************************************",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Saturday 21 March 2026  01:02:50 +0000 (0:00:00.287)       0:00:00.949 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/cinder/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating services] ************************",
                                "Saturday 21 March 2026  01:02:51 +0000 (0:00:00.870)       0:00:01.819 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 (volumev3))\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating endpoints] ***********************",
                                "Saturday 21 March 2026  01:02:54 +0000 (0:00:03.745)       0:00:05.564 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 -> https://api-int.testbed.osism.xyz:8776/v3/%(tenant_id)s -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 -> https://api.testbed.osism.xyz:8776/v3/%(tenant_id)s -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating projects] ************************",
                                "Saturday 21 March 2026  01:03:00 +0000 (0:00:05.802)       0:00:11.366 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating users] ***************************",
                                "Saturday 21 March 2026  01:03:04 +0000 (0:00:03.287)       0:00:14.654 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating roles] ***************************",
                                "Saturday 21 March 2026  01:03:07 +0000 (0:00:03.610)       0:00:18.264 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Granting user roles] **********************",
                                "Saturday 21 March 2026  01:03:10 +0000 (0:00:03.122)       0:00:21.387 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service -> admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service -> service)\u001b[0m",
                                "",
                                "TASK [cinder : Ensuring config directories exist] ******************************",
                                "Saturday 21 March 2026  01:03:17 +0000 (0:00:07.115)       0:00:28.502 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Saturday 21 March 2026  01:03:20 +0000 (0:00:02.931)       0:00:31.433 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Saturday 21 March 2026  01:03:21 +0000 (0:00:00.555)       0:00:31.989 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/cinder/tasks/external_ceph.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [cinder : Ensuring cinder service ceph config subdirs exists] *************",
                                "Saturday 21 March 2026  01:03:23 +0000 (0:00:01.670)       0:00:33.659 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder-backup)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=cinder-backup)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=cinder-backup)\u001b[0m",
                                "",
                                "TASK [cinder : Copying over multiple ceph.conf for cinder services] ************",
                                "Saturday 21 March 2026  01:03:26 +0000 (0:00:03.337)       0:00:36.996 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m",
                                "",
                                "TASK [cinder : Copy over Ceph keyring files for cinder-volume] *****************",
                                "Saturday 21 March 2026  01:03:30 +0000 (0:00:03.928)       0:00:40.924 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "",
                                "TASK [cinder : Copy over Ceph keyring files for cinder-backup] *****************",
                                "Saturday 21 March 2026  01:03:31 +0000 (0:00:01.667)       0:00:42.592 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "",
                                "TASK [cinder : Ensuring config directory has correct owner and permission] *****",
                                "Saturday 21 March 2026  01:03:34 +0000 (0:00:02.751)       0:00:45.344 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=cinder-backup)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=cinder-backup)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=cinder-backup)\u001b[0m",
                                "",
                                "TASK [cinder : Check if policies shall be overwritten] *************************",
                                "Saturday 21 March 2026  01:03:35 +0000 (0:00:01.291)       0:00:46.635 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [cinder : Set cinder policy file] *****************************************",
                                "Saturday 21 March 2026  01:03:36 +0000 (0:00:00.160)       0:00:46.796 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Saturday 21 March 2026  01:03:36 +0000 (0:00:00.415)       0:00:47.211 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/cinder/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : cinder | Copying over extra CA certificates] *********",
                                "Saturday 21 March 2026  01:03:37 +0000 (0:00:00.512)       0:00:47.724 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : cinder | Copying over backend internal TLS certificate] ***",
                                "Saturday 21 March 2026  01:03:41 +0000 (0:00:03.992)       0:00:51.717 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : cinder | Copying over backend internal TLS key] ******",
                                "Saturday 21 March 2026  01:03:41 +0000 (0:00:00.835)       0:00:52.553 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : Copying over config.json files for services] ********************",
                                "Saturday 21 March 2026  01:03:42 +0000 (0:00:01.096)       0:00:53.649 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [cinder : Copying over cinder-wsgi.conf] **********************************",
                                "Saturday 21 March 2026  01:03:47 +0000 (0:00:04.254)       0:00:57.904 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)\u001b[0m",
                                "",
                                "TASK [cinder : Copying over cinder.conf] ***************************************",
                                "Saturday 21 March 2026  01:03:49 +0000 (0:00:01.808)       0:00:59.712 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [cinder : Generating 'hostnqn' file for cinder_volume] ********************",
                                "Saturday 21 March 2026  01:04:03 +0000 (0:00:14.625)       0:01:14.338 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : Generating 'hostid' file for cinder_volume] *********************",
                                "Saturday 21 March 2026  01:04:04 +0000 (0:00:01.276)       0:01:15.614 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : Copying over existing policy file] ******************************",
                                "Saturday 21 March 2026  01:04:06 +0000 (0:00:01.214)       0:01:16.829 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : Copying over nfs_shares files for cinder_volume] ****************",
                                "Saturday 21 March 2026  01:04:07 +0000 (0:00:00.872)       0:01:17.701 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : Check cinder containers] ****************************************",
                                "Saturday 21 March 2026  01:04:07 +0000 (0:00:00.288)       0:01:17.990 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Saturday 21 March 2026  01:04:10 +0000 (0:00:02.774)       0:01:20.764 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : Creating Cinder database] ***************************************",
                                "Saturday 21 March 2026  01:04:10 +0000 (0:00:00.352)       0:01:21.116 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [cinder : Creating Cinder database user and setting permissions] **********",
                                "Saturday 21 March 2026  01:04:12 +0000 (0:00:02.065)       0:01:23.182 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [cinder : Running Cinder bootstrap container] *****************************",
                                "Saturday 21 March 2026  01:04:14 +0000 (0:00:02.321)       0:01:25.504 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [cinder : Flush handlers] *************************************************",
                                "Saturday 21 March 2026  01:04:32 +0000 (0:00:17.568)       0:01:43.073 ******** ",
                                "",
                                "TASK [cinder : Flush handlers] *************************************************",
                                "Saturday 21 March 2026  01:04:32 +0000 (0:00:00.060)       0:01:43.133 ******** ",
                                "",
                                "TASK [cinder : Flush handlers] *************************************************",
                                "Saturday 21 March 2026  01:04:32 +0000 (0:00:00.060)       0:01:43.194 ******** ",
                                "",
                                "RUNNING HANDLER [cinder : Restart cinder-api container] ************************",
                                "Saturday 21 March 2026  01:04:32 +0000 (0:00:00.060)       0:01:43.254 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [cinder : Restart cinder-scheduler container] ******************",
                                "Saturday 21 March 2026  01:04:48 +0000 (0:00:15.493)       0:01:58.748 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [cinder : Restart cinder-volume container] *********************",
                                "Saturday 21 March 2026  01:04:57 +0000 (0:00:09.775)       0:02:08.524 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [cinder : Restart cinder-backup container] *********************",
                                "Saturday 21 March 2026  01:05:14 +0000 (0:00:16.801)       0:02:25.325 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [cinder : Wait for cinder services to update service versions] ***",
                                "Saturday 21 March 2026  01:05:20 +0000 (0:00:05.377)       0:02:30.703 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=31  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  01:05:20 +0000 (0:00:00.273)       0:02:30.976 ******** ",
                                "=============================================================================== ",
                                "cinder : Running Cinder bootstrap container ---------------------------- 17.57s",
                                "cinder : Restart cinder-volume container ------------------------------- 16.80s",
                                "cinder : Restart cinder-api container ---------------------------------- 15.49s",
                                "cinder : Copying over cinder.conf -------------------------------------- 14.63s",
                                "cinder : Restart cinder-scheduler container ----------------------------- 9.78s",
                                "service-ks-register : cinder | Granting user roles ---------------------- 7.12s",
                                "service-ks-register : cinder | Creating endpoints ----------------------- 5.81s",
                                "cinder : Restart cinder-backup container -------------------------------- 5.38s",
                                "cinder : Copying over config.json files for services -------------------- 4.25s",
                                "service-cert-copy : cinder | Copying over extra CA certificates --------- 3.99s",
                                "cinder : Copying over multiple ceph.conf for cinder services ------------ 3.93s",
                                "service-ks-register : cinder | Creating services ------------------------ 3.75s",
                                "service-ks-register : cinder | Creating users --------------------------- 3.61s",
                                "cinder : Ensuring cinder service ceph config subdirs exists ------------- 3.34s",
                                "service-ks-register : cinder | Creating projects ------------------------ 3.28s",
                                "service-ks-register : cinder | Creating roles --------------------------- 3.12s",
                                "cinder : Ensuring config directories exist ------------------------------ 2.93s",
                                "cinder : Check cinder containers ---------------------------------------- 2.77s",
                                "cinder : Copy over Ceph keyring files for cinder-backup ----------------- 2.75s",
                                "cinder : Creating Cinder database user and setting permissions ---------- 2.32s",
                                "\u001b[32m2026-03-21 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:05:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 091d5fc7-25f2-481e-ad72-a993e0b38728 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3e53c3c9-52ef-4412-9e4e-44e242f3780a is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  01:04:43 +0000 (0:00:00.222)       0:00:00.222 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  01:04:43 +0000 (0:00:00.372)       0:00:00.595 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_nova_True)\u001b[0m",
                                "",
                                "PLAY [Wait for the Nova service] ***********************************************",
                                "",
                                "TASK [Waiting for Nova public port to be UP] ***********************************",
                                "Saturday 21 March 2026  01:04:44 +0000 (0:00:00.521)       0:00:01.116 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  01:06:39 +0000 (0:01:55.136)       0:01:56.253 ******** ",
                                "=============================================================================== ",
                                "Waiting for Nova public port to be UP --------------------------------- 115.13s",
                                "Group hosts based on enabled services ----------------------------------- 0.52s",
                                "Group hosts based on Kolla action --------------------------------------- 0.37s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  01:05:04 +0000 (0:00:00.309)       0:00:00.309 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  01:05:04 +0000 (0:00:00.284)       0:00:00.593 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_grafana_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_grafana_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_grafana_True)\u001b[0m",
                                "",
                                "PLAY [Apply role grafana] ******************************************************",
                                "",
                                "TASK [grafana : include_tasks] *************************************************",
                                "Saturday 21 March 2026  01:05:05 +0000 (0:00:00.303)       0:00:00.897 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/grafana/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [grafana : Ensuring config directories exist] *****************************",
                                "Saturday 21 March 2026  01:05:05 +0000 (0:00:00.491)       0:00:01.389 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "",
                                "TASK [grafana : Check if extra configuration file exists] **********************",
                                "Saturday 21 March 2026  01:05:06 +0000 (0:00:01.157)       0:00:02.547 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped '/operations/prometheus/grafana' path due to this access\u001b[0m",
                                "\u001b[1;35missue: '/operations/prometheus/grafana' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [grafana : include_tasks] *************************************************",
                                "Saturday 21 March 2026  01:05:07 +0000 (0:00:00.984)       0:00:03.531 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/grafana/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : grafana | Copying over extra CA certificates] ********",
                                "Saturday 21 March 2026  01:05:08 +0000 (0:00:00.507)       0:00:04.038 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : grafana | Copying over backend internal TLS certificate] ***",
                                "Saturday 21 March 2026  01:05:09 +0000 (0:00:01.330)       0:00:05.369 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : grafana | Copying over backend internal TLS key] *****",
                                "Saturday 21 March 2026  01:05:09 +0000 (0:00:00.372)       0:00:05.742 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Copying over config.json files] ********************************",
                                "Saturday 21 March 2026  01:05:10 +0000 (0:00:00.577)       0:00:06.319 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "",
                                "TASK [grafana : Copying over grafana.ini] **************************************",
                                "Saturday 21 March 2026  01:05:11 +0000 (0:00:01.314)       0:00:07.634 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "",
                                "TASK [grafana : Copying over extra configuration file] *************************",
                                "Saturday 21 March 2026  01:05:13 +0000 (0:00:01.189)       0:00:08.823 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Configuring Prometheus as data source for Grafana] *************",
                                "Saturday 21 March 2026  01:05:13 +0000 (0:00:00.301)       0:00:09.125 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m",
                                "",
                                "TASK [grafana : Configuring dashboards provisioning] ***************************",
                                "Saturday 21 March 2026  01:05:14 +0000 (0:00:01.086)       0:00:10.212 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m",
                                "",
                                "TASK [grafana : Find custom grafana dashboards] ********************************",
                                "Saturday 21 March 2026  01:05:15 +0000 (0:00:01.061)       0:00:11.273 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [grafana : Find templated grafana dashboards] *****************************",
                                "Saturday 21 March 2026  01:05:16 +0000 (0:00:01.041)       0:00:12.315 ******** ",
                                "\u001b[1;35m[WARNING]: Skipped '/etc/kolla/grafana/dashboards' path due to this access\u001b[0m",
                                "\u001b[1;35missue: '/etc/kolla/grafana/dashboards' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Prune templated Grafana dashboards] ****************************",
                                "Saturday 21 March 2026  01:05:17 +0000 (0:00:00.657)       0:00:12.972 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Copying over custom dashboards] ********************************",
                                "Saturday 21 March 2026  01:05:17 +0000 (0:00:00.304)       0:00:13.277 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 121701, 'inode': 1106291, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9733195, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 121701, 'inode': 1106291, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9733195, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 121701, 'inode': 1106291, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9733195, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/cephfsdashboard.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfsdashboard.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 143913, 'inode': 1106321, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.985684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/cephfsdashboard.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfsdashboard.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 143913, 'inode': 1106321, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.985684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/cephfsdashboard.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfsdashboard.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 143913, 'inode': 1106321, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.985684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26019, 'inode': 1106342, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9952614, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26019, 'inode': 1106342, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9952614, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26019, 'inode': 1106342, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9952614, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1106319, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.984684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1106319, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.984684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1106319, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.984684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 170293, 'inode': 1106343, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.996892, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 170293, 'inode': 1106343, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.996892, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 170293, 'inode': 1106343, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.996892, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph-nvmeof-performance.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-nvmeof-performance.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 33297, 'inode': 1106305, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.978684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph-nvmeof-performance.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-nvmeof-performance.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 33297, 'inode': 1106305, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.978684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph-nvmeof-performance.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-nvmeof-performance.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 33297, 'inode': 1106305, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.978684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26346, 'inode': 1106330, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.989684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26346, 'inode': 1106330, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.989684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26346, 'inode': 1106330, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.989684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 46110, 'inode': 1106339, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9926841, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 46110, 'inode': 1106339, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9926841, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 46110, 'inode': 1106339, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9926841, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1106287, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.971438, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1106287, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.971438, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1106287, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.971438, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1106300, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9748197, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1106300, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9748197, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1106300, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9748197, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1106320, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.984684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1106320, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.984684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1106320, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.984684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19231, 'inode': 1106334, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9914844, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19231, 'inode': 1106334, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9914844, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19231, 'inode': 1106334, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9914844, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13320, 'inode': 1106341, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9948537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13320, 'inode': 1106341, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9948537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13320, 'inode': 1106341, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9948537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1106316, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.982684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1106316, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.982684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1106316, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.982684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 20042, 'inode': 1106338, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9926841, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 20042, 'inode': 1106338, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9926841, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 20042, 'inode': 1106338, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9926841, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/smb-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/smb-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29877, 'inode': 1106347, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9976842, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/smb-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/smb-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29877, 'inode': 1106347, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9976842, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/smb-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/smb-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29877, 'inode': 1106347, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9976842, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38375, 'inode': 1106331, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.990684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38375, 'inode': 1106331, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.990684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38375, 'inode': 1106331, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.990684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 63043, 'inode': 1106328, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9889889, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 63043, 'inode': 1106328, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9889889, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 63043, 'inode': 1106328, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9889889, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27387, 'inode': 1106324, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9882722, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27387, 'inode': 1106324, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9882722, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27387, 'inode': 1106324, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9882722, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49016, 'inode': 1106336, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.991684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49016, 'inode': 1106336, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.991684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49016, 'inode': 1106336, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.991684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 43303, 'inode': 1106322, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.986684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 43303, 'inode': 1106322, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.986684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 43303, 'inode': 1106322, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.986684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16614, 'inode': 1106340, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.993684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16614, 'inode': 1106340, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.993684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16614, 'inode': 1106340, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.993684, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph-nvmeof.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-nvmeof.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 52667, 'inode': 1106314, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9806838, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph-nvmeof.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-nvmeof.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 52667, 'inode': 1106314, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9806838, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph-nvmeof.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-nvmeof.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 52667, 'inode': 1106314, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9806838, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1106481, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.044685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1106481, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.044685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1106481, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.044685, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1106414, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0217268, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1106414, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0217268, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1106414, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0217268, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1106393, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.014711, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1106393, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.014711, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1106393, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.014711, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15767, 'inode': 1106442, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0248306, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15767, 'inode': 1106442, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0248306, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15767, 'inode': 1106442, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0248306, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1106350, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9986842, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1106350, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9986842, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1106350, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9986842, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1106461, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0348463, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1106461, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0348463, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1106461, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0348463, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1106444, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0326846, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1106444, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0326846, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1106444, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0326846, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22303, 'inode': 1106462, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0355852, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22303, 'inode': 1106462, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0355852, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22303, 'inode': 1106462, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0355852, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1106476, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0416849, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1106476, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0416849, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1106476, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0416849, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21194, 'inode': 1106459, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0336847, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21194, 'inode': 1106459, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0336847, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21194, 'inode': 1106459, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0336847, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1106432, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0229754, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1106432, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0229754, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1106432, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0229754, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1106409, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0184097, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1106409, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0184097, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1106409, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0184097, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1106426, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0226433, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1106426, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0226433, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1106426, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0226433, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1106398, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0175004, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1106398, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0175004, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1106398, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0175004, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15957, 'inode': 1106437, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0236845, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15957, 'inode': 1106437, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0236845, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15957, 'inode': 1106437, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0236845, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1106469, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0406847, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1106469, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0406847, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1106469, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0406847, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1106466, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0376847, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1106466, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0376847, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1106466, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0376847, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1106351, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9995236, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1106351, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9995236, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1106351, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052232.9995236, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1106352, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0139942, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1106352, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0139942, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1106352, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0139942, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1106457, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0336847, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1106457, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0336847, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1106457, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0336847, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21951, 'inode': 1106464, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0366848, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21951, 'inode': 1106464, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0366848, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21951, 'inode': 1106464, 'dev': 108, 'nlink': 1, 'atime': 1774051353.0, 'mtime': 1774051353.0, 'ctime': 1774052233.0366848, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "",
                                "TASK [grafana : Check grafana containers] **************************************",
                                "Saturday 21 March 2026  01:05:52 +0000 (0:00:34.612)       0:00:47.889 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "",
                                "TASK [grafana : Creating grafana database] *************************************",
                                "Saturday 21 March 2026  01:05:53 +0000 (0:00:00.992)       0:00:48.882 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [grafana : Creating grafana database user and setting permissions] ********",
                                "Saturday 21 March 2026  01:05:55 +0000 (0:00:01.902)       0:00:50.784 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [grafana : Flush handlers] ************************************************",
                                "Saturday 21 March 2026  01:05:56 +0000 (0:00:01.794)       0:00:52.579 ******** ",
                                "",
                                "TASK [grafana : Flush handlers] ************************************************",
                                "Saturday 21 March 2026  01:05:56 +0000 (0:00:00.059)       0:00:52.639 ******** ",
                                "",
                                "TASK [grafana : Flush handlers] ************************************************",
                                "Saturday 21 March 2026  01:05:56 +0000 (0:00:00.064)       0:00:52.703 ******** ",
                                "",
                                "RUNNING HANDLER [grafana : Restart first grafana container] ********************",
                                "Saturday 21 March 2026  01:05:56 +0000 (0:00:00.065)       0:00:52.768 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [grafana : Waiting for grafana to start on first node] *********",
                                "Saturday 21 March 2026  01:06:03 +0000 (0:00:06.619)       0:00:59.388 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (12 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [grafana : Restart remaining grafana containers] ***************",
                                "Saturday 21 March 2026  01:06:17 +0000 (0:00:13.744)       0:01:13.133 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [grafana : Wait for grafana application ready] ****************************",
                                "Saturday 21 March 2026  01:06:37 +0000 (0:00:19.766)       0:01:32.900 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [grafana : Remove old grafana docker volume] ******************************",
                                "Saturday 21 March 2026  01:06:39 +0000 (0:00:01.914)       0:01:34.814 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Enable grafana datasources] ************************************",
                                "Saturday 21 March 2026  01:06:39 +0000 (0:00:00.263)       0:01:35.077 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'influxdb', 'value': {'enabled': False, 'data': {'isDefault': True, 'database': 'telegraf', 'name': 'telegraf', 'type': 'influxdb', 'url': 'https://api-int.testbed.osism.xyz:8086', 'access': 'proxy', 'basicAuth': False}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'enabled': True, 'data': {'name': 'opensearch', 'type': 'grafana-opensearch-datasource', 'access': 'proxy', 'url': 'https://api-int.testbed.osism.xyz:9200', 'jsonData': {'flavor': 'OpenSearch', 'database': 'flog-*', 'version': '2.11.1', 'timeField': '@timestamp', 'logLevelField': 'log_level'}}}})\u001b[0m",
                                "",
                                "TASK [grafana : Disable Getting Started panel] *********************************",
                                "Saturday 21 March 2026  01:06:41 +0000 (0:00:02.287)       0:01:37.364 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  01:06:41 +0000 (0:00:00.262)       0:01:37.626 ******** ",
                                "=============================================================================== ",
                                "grafana : Copying over custom dashboards ------------------------------- 34.61s",
                                "grafana : Restart remaining grafana containers ------------------------- 19.77s",
                                "grafana : Waiting for grafana to start on first node ------------------- 13.74s",
                                "grafana : Restart first grafana container ------------------------------- 6.62s",
                                "grafana : Enable grafana datasources ------------------------------------ 2.29s",
                                "grafana : Wait for grafana application ready ---------------------------- 1.91s",
                                "grafana : Creating grafana database ------------------------------------- 1.90s",
                                "grafana : Creating grafana database user and setting permissions -------- 1.79s",
                                "service-cert-copy : grafana | Copying over extra CA certificates -------- 1.33s",
                                "grafana : Copying over config.json files -------------------------------- 1.31s",
                                "grafana : Copying over grafana.ini -------------------------------------- 1.19s",
                                "grafana : Ensuring config directories exist ----------------------------- 1.16s",
                                "grafana : Configuring Prometheus as data source for Grafana ------------- 1.09s",
                                "grafana : Configuring dashboards provisioning --------------------------- 1.06s",
                                "grafana : Find custom grafana dashboards -------------------------------- 1.04s",
                                "grafana : Check grafana containers -------------------------------------- 0.99s",
                                "grafana : Check if extra configuration file exists ---------------------- 0.98s",
                                "grafana : Find templated grafana dashboards ----------------------------- 0.66s",
                                "service-cert-copy : grafana | Copying over backend internal TLS key ----- 0.58s",
                                "grafana : include_tasks ------------------------------------------------- 0.51s",
                                "\u001b[32m2026-03-21 01:06:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:06:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:07:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:08:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:09:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1373aa5e-75c8-4d0f-adef-c160cfd35131 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on OpenStack release] **********************************",
                                "Saturday 21 March 2026  01:02:57 +0000 (0:00:00.278)       0:00:00.278 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  01:02:58 +0000 (0:00:00.534)       0:00:00.812 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  01:02:59 +0000 (0:00:00.736)       0:00:01.548 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=enable_nova_True)\u001b[0m",
                                "",
                                "PLAY [Bootstrap nova API databases] ********************************************",
                                "",
                                "TASK [Bootstrap deploy] ********************************************************",
                                "Saturday 21 March 2026  01:03:00 +0000 (0:00:01.137)       0:00:02.686 ******** ",
                                "\u001b[0;36mincluded: nova for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova : Creating Nova databases] ******************************************",
                                "Saturday 21 March 2026  01:03:01 +0000 (0:00:01.009)       0:00:03.696 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova_cell0)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova_api)\u001b[0m",
                                "",
                                "TASK [nova : Creating Nova databases user and setting permissions] *************",
                                "Saturday 21 March 2026  01:03:05 +0000 (0:00:04.346)       0:00:08.042 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Ensuring config directories exist] ********************************",
                                "Saturday 21 March 2026  01:03:09 +0000 (0:00:04.038)       0:00:12.081 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Copying over config.json files for nova-api-bootstrap] ************",
                                "Saturday 21 March 2026  01:03:10 +0000 (0:00:00.673)       0:00:12.754 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Copying over nova.conf for nova-api-bootstrap] ********************",
                                "Saturday 21 March 2026  01:03:11 +0000 (0:00:01.338)       0:00:14.093 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : include_tasks] ****************************************************",
                                "Saturday 21 March 2026  01:03:14 +0000 (0:00:02.956)       0:00:17.049 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Running Nova API bootstrap container] *****************************",
                                "Saturday 21 March 2026  01:03:15 +0000 (0:00:00.412)       0:00:17.462 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Create cell0 mappings] ********************************************",
                                "Saturday 21 March 2026  01:03:44 +0000 (0:00:29.554)       0:00:47.016 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Get a list of existing cells] ********************************",
                                "Saturday 21 March 2026  01:03:58 +0000 (0:00:14.206)       0:01:01.222 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract current cell settings from list] *********************",
                                "Saturday 21 March 2026  01:04:09 +0000 (0:00:10.978)       0:01:12.201 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Update cell0 mappings] ********************************************",
                                "Saturday 21 March 2026  01:04:10 +0000 (0:00:00.786)       0:01:12.988 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : include_tasks] ****************************************************",
                                "Saturday 21 March 2026  01:04:10 +0000 (0:00:00.412)       0:01:13.400 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova/tasks/bootstrap_service.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova : Running Nova API bootstrap container] *****************************",
                                "Saturday 21 March 2026  01:04:11 +0000 (0:00:00.721)       0:01:14.121 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Bootstrap upgrade] *******************************************************",
                                "Saturday 21 March 2026  01:04:29 +0000 (0:00:18.095)       0:01:32.216 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Bootstrap nova cell databases] *******************************************",
                                "",
                                "TASK [Bootstrap deploy] ********************************************************",
                                "Saturday 21 March 2026  01:04:30 +0000 (0:00:00.288)       0:01:32.505 ******** ",
                                "\u001b[0;36mincluded: nova-cell for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova-cell : Creating Nova cell database] *********************************",
                                "Saturday 21 March 2026  01:04:30 +0000 (0:00:00.629)       0:01:33.135 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Creating Nova cell database user and setting permissions] ****",
                                "Saturday 21 March 2026  01:04:32 +0000 (0:00:01.747)       0:01:34.883 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-rabbitmq : nova | Ensure RabbitMQ vhosts exist] ******************",
                                "Saturday 21 March 2026  01:04:34 +0000 (0:00:01.739)       0:01:36.622 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-rabbitmq : nova | Ensure RabbitMQ users exist] *******************",
                                "Saturday 21 March 2026  01:04:34 +0000 (0:00:00.407)       0:01:37.029 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> {{ service_rabbitmq_delegate_host }}]\u001b[0m",
                                "",
                                "TASK [service-rabbitmq : nova | Ensure RabbitMQ vhosts exist] ******************",
                                "Saturday 21 March 2026  01:04:40 +0000 (0:00:06.230)       0:01:43.259 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-rabbitmq : nova | Ensure RabbitMQ users exist] *******************",
                                "Saturday 21 March 2026  01:04:41 +0000 (0:00:00.507)       0:01:43.767 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Ensuring config directories exist] ***************************",
                                "Saturday 21 March 2026  01:04:42 +0000 (0:00:01.435)       0:01:45.203 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over config.json files for nova-cell-bootstrap] ******",
                                "Saturday 21 March 2026  01:04:43 +0000 (0:00:00.479)       0:01:45.682 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over nova.conf for nova-cell-bootstrap] **************",
                                "Saturday 21 March 2026  01:04:44 +0000 (0:00:00.921)       0:01:46.604 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Running Nova cell bootstrap container] ***********************",
                                "Saturday 21 March 2026  01:04:46 +0000 (0:00:02.000)       0:01:48.604 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Get a list of existing cells] ********************************",
                                "Saturday 21 March 2026  01:05:06 +0000 (0:00:20.161)       0:02:08.766 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract current cell settings from list] *********************",
                                "Saturday 21 March 2026  01:05:18 +0000 (0:00:11.738)       0:02:20.504 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Create cell] *************************************************",
                                "Saturday 21 March 2026  01:05:19 +0000 (0:00:01.029)       0:02:21.534 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Update cell] *************************************************",
                                "Saturday 21 March 2026  01:05:32 +0000 (0:00:13.315)       0:02:34.850 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Bootstrap upgrade] *******************************************************",
                                "Saturday 21 March 2026  01:05:33 +0000 (0:00:01.107)       0:02:35.957 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role nova] *********************************************************",
                                "",
                                "TASK [nova : include_tasks] ****************************************************",
                                "Saturday 21 March 2026  01:05:33 +0000 (0:00:00.265)       0:02:36.223 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating services] **************************",
                                "Saturday 21 March 2026  01:05:34 +0000 (0:00:00.562)       0:02:36.786 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy (compute_legacy)) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova (compute))\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating endpoints] *************************",
                                "Saturday 21 March 2026  01:05:37 +0000 (0:00:03.189)       0:02:39.976 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy -> https://api-int.testbed.osism.xyz:8774/v2/%(tenant_id)s -> internal) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy -> https://api.testbed.osism.xyz:8774/v2/%(tenant_id)s -> public) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> https://api-int.testbed.osism.xyz:8774/v2.1 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> https://api.testbed.osism.xyz:8774/v2.1 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating projects] **************************",
                                "Saturday 21 March 2026  01:05:43 +0000 (0:00:06.005)       0:02:45.981 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating users] *****************************",
                                "Saturday 21 March 2026  01:05:46 +0000 (0:00:02.600)       0:02:48.581 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating roles] *****************************",
                                "Saturday 21 March 2026  01:05:49 +0000 (0:00:03.382)       0:02:51.964 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Granting user roles] ************************",
                                "Saturday 21 March 2026  01:05:52 +0000 (0:00:02.783)       0:02:54.748 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service -> admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service -> service)\u001b[0m",
                                "",
                                "TASK [nova : Ensuring config directories exist] ********************************",
                                "Saturday 21 March 2026  01:05:58 +0000 (0:00:05.976)       0:03:00.725 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova : Check if policies shall be overwritten] ***************************",
                                "Saturday 21 March 2026  01:06:00 +0000 (0:00:01.761)       0:03:02.487 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Set nova policy file] *********************************************",
                                "Saturday 21 March 2026  01:06:00 +0000 (0:00:00.115)       0:03:02.603 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Check for vendordata file] ****************************************",
                                "Saturday 21 March 2026  01:06:00 +0000 (0:00:00.311)       0:03:02.914 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [nova : Set vendordata file path] *****************************************",
                                "Saturday 21 March 2026  01:06:01 +0000 (0:00:00.737)       0:03:03.651 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : include_tasks] ****************************************************",
                                "Saturday 21 March 2026  01:06:01 +0000 (0:00:00.286)       0:03:03.937 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over extra CA certificates] ***********",
                                "Saturday 21 March 2026  01:06:02 +0000 (0:00:00.730)       0:03:04.668 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over backend internal TLS certificate] ***",
                                "Saturday 21 March 2026  01:06:04 +0000 (0:00:02.001)       0:03:06.669 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over backend internal TLS key] ********",
                                "Saturday 21 March 2026  01:06:04 +0000 (0:00:00.570)       0:03:07.240 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Copying over config.json files for services] **********************",
                                "Saturday 21 March 2026  01:06:05 +0000 (0:00:00.923)       0:03:08.164 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova : Copying over nova.conf] *******************************************",
                                "Saturday 21 March 2026  01:06:07 +0000 (0:00:02.111)       0:03:10.276 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova : Copying over existing policy file] ********************************",
                                "Saturday 21 March 2026  01:06:13 +0000 (0:00:05.140)       0:03:15.416 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Copying over nova-api-wsgi.conf] **********************************",
                                "Saturday 21 March 2026  01:06:13 +0000 (0:00:00.591)       0:03:16.008 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Copying over vendordata file] *************************************",
                                "Saturday 21 March 2026  01:06:15 +0000 (0:00:01.643)       0:03:17.652 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Check nova containers] ********************************************",
                                "Saturday 21 March 2026  01:06:15 +0000 (0:00:00.306)       0:03:17.959 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova : Flush handlers] ***************************************************",
                                "Saturday 21 March 2026  01:06:17 +0000 (0:00:01.688)       0:03:19.647 ******** ",
                                "",
                                "TASK [nova : Flush handlers] ***************************************************",
                                "Saturday 21 March 2026  01:06:17 +0000 (0:00:00.144)       0:03:19.792 ******** ",
                                "",
                                "TASK [nova : Flush handlers] ***************************************************",
                                "Saturday 21 March 2026  01:06:17 +0000 (0:00:00.131)       0:03:19.923 ******** ",
                                "",
                                "RUNNING HANDLER [nova : Restart nova-scheduler container] **********************",
                                "Saturday 21 March 2026  01:06:17 +0000 (0:00:00.302)       0:03:20.226 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova : Restart nova-api container] ****************************",
                                "Saturday 21 March 2026  01:06:31 +0000 (0:00:13.745)       0:03:33.971 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Apply role nova-cell] ****************************************************",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Saturday 21 March 2026  01:06:37 +0000 (0:00:05.777)       0:03:39.749 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Saturday 21 March 2026  01:06:38 +0000 (0:00:01.178)       0:03:40.927 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Load and persist br_netfilter module] ************************************",
                                "Saturday 21 March 2026  01:06:39 +0000 (0:00:00.706)       0:03:41.634 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: module-load for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [module-load : Load modules] **********************************************",
                                "Saturday 21 March 2026  01:06:40 +0000 (0:00:00.800)       0:03:42.434 ******** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=br_netfilter)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=br_netfilter)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=br_netfilter)\u001b[0m",
                                "",
                                "TASK [module-load : Persist modules via modules-load.d] ************************",
                                "Saturday 21 March 2026  01:06:41 +0000 (0:00:01.205)       0:03:43.639 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=br_netfilter)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=br_netfilter)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=br_netfilter)\u001b[0m",
                                "",
                                "TASK [module-load : Drop module persistence] ***********************************",
                                "Saturday 21 March 2026  01:06:42 +0000 (0:00:01.110)       0:03:44.750 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=br_netfilter) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=br_netfilter) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=br_netfilter) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Enable bridge-nf-call sysctl variables] **********************",
                                "Saturday 21 March 2026  01:06:42 +0000 (0:00:00.519)       0:03:45.270 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m",
                                "",
                                "TASK [nova-cell : Install udev kolla kvm rules] ********************************",
                                "Saturday 21 March 2026  01:06:44 +0000 (0:00:02.097)       0:03:47.367 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Mask qemu-kvm service] ***************************************",
                                "Saturday 21 March 2026  01:06:45 +0000 (0:00:01.006)       0:03:48.373 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [nova-cell : Ensuring config directories exist] ***************************",
                                "Saturday 21 March 2026  01:06:47 +0000 (0:00:01.601)       0:03:49.975 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Saturday 21 March 2026  01:06:49 +0000 (0:00:02.252)       0:03:52.228 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/copy-certs.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over extra CA certificates] ***********",
                                "Saturday 21 March 2026  01:06:50 +0000 (0:00:01.013)       0:03:53.241 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over backend internal TLS certificate] ***",
                                "Saturday 21 March 2026  01:06:53 +0000 (0:00:03.140)       0:03:56.382 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over backend internal TLS key] ********",
                                "Saturday 21 March 2026  01:06:55 +0000 (0:00:01.439)       0:03:57.821 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvir\u001b[32m2026-03-21 01:10:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "td:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Saturday 21 March 2026  01:06:57 +0000 (0:00:02.005)       0:03:59.827 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/external_ceph.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [nova-cell : Check nova keyring file] *************************************",
                                "Saturday 21 March 2026  01:06:58 +0000 (0:00:00.979)       0:04:00.807 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [nova-cell : Check cinder keyring file] ***********************************",
                                "Saturday 21 March 2026  01:06:59 +0000 (0:00:00.847)       0:04:01.654 ******** ",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract nova key from file] **********************************",
                                "Saturday 21 March 2026  01:07:00 +0000 (0:00:00.951)       0:04:02.606 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract cinder key from file] ********************************",
                                "Saturday 21 March 2026  01:07:00 +0000 (0:00:00.509)       0:04:03.115 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copy over ceph nova keyring file] ****************************",
                                "Saturday 21 March 2026  01:07:01 +0000 (0:00:00.458)       0:04:03.574 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m",
                                "",
                                "TASK [nova-cell : Copy over ceph cinder keyring file] **************************",
                                "Saturday 21 March 2026  01:07:02 +0000 (0:00:00.981)       0:04:04.555 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m",
                                "",
                                "TASK [nova-cell : Copy over ceph.conf] *****************************************",
                                "Saturday 21 March 2026  01:07:03 +0000 (0:00:01.137)       0:04:05.693 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=nova-libvirt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=nova-libvirt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=nova-libvirt)\u001b[0m",
                                "",
                                "TASK [nova-cell : Ensure /etc/ceph directory exists (host libvirt)] ************",
                                "Saturday 21 March 2026  01:07:06 +0000 (0:00:03.282)       0:04:08.975 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copy over ceph.conf (host libvirt)] **************************",
                                "Saturday 21 March 2026  01:07:06 +0000 (0:00:00.288)       0:04:09.264 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Ensuring libvirt secrets directory exists] *******************",
                                "Saturday 21 March 2026  01:07:07 +0000 (0:00:00.284)       0:04:09.549 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Pushing nova secret xml for libvirt] *************************",
                                "Saturday 21 March 2026  01:07:08 +0000 (0:00:01.433)       0:04:10.982 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'})\u001b[0m",
                                "",
                                "TASK [nova-cell : Pushing secrets key for libvirt] *****************************",
                                "Saturday 21 March 2026  01:07:11 +0000 (0:00:03.333)       0:04:14.316 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Include tasks from qemu_wrapper.yml] *************************",
                                "Saturday 21 March 2026  01:07:14 +0000 (0:00:03.032)       0:04:17.349 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/qemu_wrapper.yml for testbed-node-4, testbed-node-3, testbed-node-5\u001b[0m",
                                "",
                                "TASK [nova-cell : Check qemu wrapper file] *************************************",
                                "Saturday 21 March 2026  01:07:16 +0000 (0:00:01.743)       0:04:19.093 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copy qemu wrapper] *******************************************",
                                "Saturday 21 March 2026  01:07:17 +0000 (0:00:01.070)       0:04:20.163 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Check if policies shall be overwritten] **********************",
                                "Saturday 21 March 2026  01:07:18 +0000 (0:00:00.323)       0:04:20.487 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [nova-cell : Set nova policy file] ****************************************",
                                "Saturday 21 March 2026  01:07:18 +0000 (0:00:00.124)       0:04:20.611 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Check for vendordata file] ***********************************",
                                "Saturday 21 March 2026  01:07:18 +0000 (0:00:00.772)       0:04:21.384 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "",
                                "TASK [nova-cell : Set vendordata file path] ************************************",
                                "Saturday 21 March 2026  01:07:19 +0000 (0:00:00.737)       0:04:22.121 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over config.json files for services] *****************",
                                "Saturday 21 March 2026  01:07:20 +0000 (0:00:00.556)       0:04:22.678 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over nova.conf] **************************************",
                                "Saturday 21 March 2026  01:07:23 +0000 (0:00:03.684)       0:04:26.362 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over Nova compute provider config] *******************",
                                "Saturday 21 March 2026  01:07:29 +0000 (0:00:05.753)       0:04:32.116 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over libvirt configuration] **************************",
                                "Saturday 21 March 2026  01:07:31 +0000 (0:00:01.650)       0:04:33.766 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over libvirt TLS keys] *******************************",
                                "Saturday 21 March 2026  01:07:34 +0000 (0:00:03.479)       0:04:37.246 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over libvirt SASL configuration] *********************",
                                "Saturday 21 March 2026  01:07:35 +0000 (0:00:00.732)       0:04:37.979 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying files for nova-ssh] **********************************",
                                "Saturday 21 March 2026  01:07:40 +0000 (0:00:05.150)       0:04:43.129 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying VMware vCenter CA file] ******************************",
                                "Saturday 21 March 2026  01:07:46 +0000 (0:00:06.221)       0:04:49.351 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying 'release' file for nova_compute] *********************",
                                "Saturday 21 March 2026  01:07:47 +0000 (0:00:00.557)       0:04:49.908 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Generating 'hostnqn' file for nova_compute] ******************",
                                "Saturday 21 March 2026  01:07:48 +0000 (0:00:00.796)       0:04:50.705 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Generating 'hostid' file for nova_compute] *******************",
                                "Saturday 21 March 2026  01:07:50 +0000 (0:00:01.858)       0:04:52.563 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over existing policy file] ***************************",
                                "Saturday 21 March 2026  01:07:52 +0000 (0:00:02.090)       0:04:54.654 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over vendordata file to containers] ******************",
                                "Saturday 21 March 2026  01:07:53 +0000 (0:00:01.328)       0:04:55.983 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Check nova-cell containers] **********************************",
                                "Saturday 21 March 2026  01:07:54 +0000 (0:00:00.784)       0:04:56.767 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Saturday 21 March 2026  01:07:56 +0000 (0:00:02.537)       0:04:59.305 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Saturday 21 March 2026  01:07:57 +0000 (0:00:00.761)       0:05:00.067 ******** ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Saturday 21 March 2026  01:07:57 +0000 (0:00:00.128)       0:05:00.195 ******** ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Saturday 21 March 2026  01:07:57 +0000 (0:00:00.127)       0:05:00.322 ******** ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Saturday 21 March 2026  01:07:58 +0000 (0:00:00.143)       0:05:00.466 ******** ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Saturday 21 March 2026  01:07:58 +0000 (0:00:00.139)       0:05:00.605 ******** ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Saturday 21 March 2026  01:07:58 +0000 (0:00:00.277)       0:05:00.883 ******** ",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-conductor container] *****************",
                                "Saturday 21 March 2026  01:07:58 +0000 (0:00:00.144)       0:05:01.028 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-novncproxy container] ****************",
                                "Saturday 21 March 2026  01:08:05 +0000 (0:00:06.765)       0:05:07.794 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-ssh container] ***********************",
                                "Saturday 21 March 2026  01:08:21 +0000 (0:00:16.178)       0:05:23.973 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-libvirt container] *******************",
                                "Saturday 21 March 2026  01:08:40 +0000 (0:00:19.195)       0:05:43.169 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Checking libvirt container is ready] **************",
                                "Saturday 21 March 2026  01:09:05 +0000 (0:00:24.643)       0:06:07.812 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-3]: Checking libvirt container is ready (10 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-4]: Checking libvirt container is ready (10 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-5]: Checking libvirt container is ready (10 retries left).\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Create libvirt SASL user] *************************",
                                "Saturday 21 March 2026  01:09:11 +0000 (0:00:06.091)       0:06:13.904 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-compute container] *******************",
                                "Saturday 21 March 2026  01:09:12 +0000 (0:00:00.718)       0:06:14.622 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Wait for nova-compute services to update service versions] ***",
                                "Saturday 21 March 2026  01:09:34 +0000 (0:00:21.795)       0:06:36.418 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [nova-cell : Waiting for nova-compute services to register themselves] ****",
                                "Saturday 21 March 2026  01:09:34 +0000 (0:00:00.237)       0:06:36.655 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-3 -> testbed-node-0]: Waiting for nova-compute services to register themselves (20 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [nova-cell : Fail if nova-compute service failed to register] *************",
                                "Saturday 21 March 2026  01:09:54 +0000 (0:00:20.641)       0:06:57.296 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [nova-cell : Include discover_computes.yml] *******************************",
                                "Saturday 21 March 2026  01:10:03 +0000 (0:00:08.171)       0:07:05.468 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/discover_computes.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [nova-cell : Get a list of existing cells] ********************************",
                                "Saturday 21 March 2026  01:10:06 +0000 (0:00:03.781)       0:07:09.249 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract current cell settings from list] *********************",
                                "Saturday 21 March 2026  01:10:20 +0000 (0:00:14.062)       0:07:23.312 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [nova-cell : Fail if cell settings not found] *****************************",
                                "Saturday 21 March 2026  01:10:22 +0000 (0:00:01.490)       0:07:24.803 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [nova-cell : Discover nova hosts] *****************************************",
                                "Saturday 21 March 2026  01:10:23 +0000 (0:00:01.339)       0:07:26.142 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [nova-cell : Remove old nova_libvirt_secrets container volume] ************",
                                "Saturday 21 March 2026  01:10:36 +0000 (0:00:12.809)       0:07:38.952 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Refresh nova scheduler cell cache] ***************************************",
                                "",
                                "TASK [nova : Refresh cell cache in nova scheduler] *****************************",
                                "Saturday 21 March 2026  01:10:38 +0000 (0:00:01.930)       0:07:40.883 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Reload global Nova super conductor services] *****************************",
                                "",
                                "TASK [nova : Reload nova super conductor services to remove RPC version pin] ***",
                                "Saturday 21 March 2026  01:10:39 +0000 (0:00:01.021)       0:07:41.904 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Reload Nova cell services] ***********************************************",
                                "",
                                "TASK [nova-cell : Reload nova cell services to remove RPC version cap] *********",
                                "Saturday 21 March 2026  01:10:39 +0000 (0:00:00.485)       0:07:42.389 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Reload global Nova API services] *****************************************",
                                "",
                                "TASK [nova : Reload nova API services to remove RPC version pin] ***************",
                                "Saturday 21 March 2026  01:10:41 +0000 (0:00:01.266)       0:07:43.656 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-scheduler) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-api) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-scheduler) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-api) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-scheduler) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-api) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Run Nova API online data migrations] *************************************",
                                "",
                                "TASK [nova : Run Nova API online database migrations] **************************",
                                "Saturday 21 March 2026  01:10:41 +0000 (0:00:00.668)       0:07:44.324 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Run Nova cell online data migrations] ************************************",
                                "",
                                "TASK [nova-cell : Run Nova cell online database migrations] ********************",
                                "Saturday 21 March 2026  01:10:42 +0000 (0:00:00.629)       0:07:44.954 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=54  \u001b[0m \u001b[0;33mchanged=35  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=46  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=53  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=53  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=46  \u001b[0m \u001b[0;33mchanged=28  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=21  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=40  \u001b[0m \u001b[0;33mchanged=28  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=40  \u001b[0m \u001b[0;33mchanged=28  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  01:10:42 +0000 (0:00:00.411)       0:07:45.366 ******** ",
                                "=============================================================================== ",
                                "nova : Running Nova API bootstrap container ---------------------------- 29.55s",
                                "nova-cell : Restart nova-libvirt container ----------------------------- 24.64s",
                                "nova-cell : Restart nova-compute container ----------------------------- 21.80s",
                                "nova-cell : Waiting for nova-compute services to register themselves --- 20.64s",
                                "nova-cell : Running Nova cell bootstrap container ---------------------- 20.16s",
                                "nova-cell : Restart nova-ssh container --------------------------------- 19.20s",
                                "nova : Running Nova API bootstrap container ---------------------------- 18.10s",
                                "nova-cell : Restart nova-novncproxy container -------------------------- 16.18s",
                                "nova : Create cell0 mappings ------------------------------------------- 14.21s",
                                "nova-cell : Get a list of existing cells ------------------------------- 14.06s",
                                "nova : Restart nova-scheduler container -------------------------------- 13.75s",
                                "nova-cell : Create cell ------------------------------------------------ 13.32s",
                                "nova-cell : Discover nova hosts ---------------------------------------- 12.81s",
                                "nova-cell : Get a list of existing cells ------------------------------- 11.74s",
                                "nova-cell : Get a list of existing cells ------------------------------- 10.98s",
                                "nova-cell : Fail if nova-compute service failed to register ------------- 8.17s",
                                "nova-cell : Restart nova-conductor container ---------------------------- 6.77s",
                                "service-rabbitmq : nova | Ensure RabbitMQ users exist ------------------- 6.23s",
                                "nova-cell : Copying files for nova-ssh ---------------------------------- 6.22s",
                                "nova-cell : Checking libvirt container is ready ------------------------- 6.09s",
                                "\u001b[32m2026-03-21 01:10:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:10:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8bbb4571-cca3-452f-b444-abd85595f002 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  01:06:43 +0000 (0:00:00.312)       0:00:00.312 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  01:06:43 +0000 (0:00:00.281)       0:00:00.594 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_octavia_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_octavia_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_octavia_True)\u001b[0m",
                                "",
                                "PLAY [Apply role octavia] ******************************************************",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Saturday 21 March 2026  01:06:43 +0000 (0:00:00.294)       0:00:00.888 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating services] ***********************",
                                "Saturday 21 March 2026  01:06:44 +0000 (0:00:00.622)       0:00:01.510 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia (load-balancer))\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating endpoints] **********************",
                                "Saturday 21 March 2026  01:06:47 +0000 (0:00:03.691)       0:00:05.202 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> https://api-int.testbed.osism.xyz:9876 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> https://api.testbed.osism.xyz:9876 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating projects] ***********************",
                                "Saturday 21 March 2026  01:06:54 +0000 (0:00:06.475)       0:00:11.678 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating users] **************************",
                                "Saturday 21 March 2026  01:06:57 +0000 (0:00:02.795)       0:00:14.473 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating roles] **************************",
                                "Saturday 21 March 2026  01:07:04 +0000 (0:00:07.483)       0:00:21.957 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Granting user roles] *********************",
                                "Saturday 21 March 2026  01:07:08 +0000 (0:00:03.400)       0:00:25.357 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service -> admin)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=octavia -> service -> admin)\u001b[0m",
                                "",
                                "TASK [octavia : Adding octavia related roles] **********************************",
                                "Saturday 21 March 2026  01:07:16 +0000 (0:00:07.991)       0:00:33.348 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_observer)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_global_observer)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_member)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_quota_admin)\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Saturday 21 March 2026  01:07:30 +0000 (0:00:14.921)       0:00:48.270 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/prepare.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [octavia : Create amphora flavor] *****************************************",
                                "Saturday 21 March 2026  01:07:31 +0000 (0:00:00.836)       0:00:49.107 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Create nova keypair for amphora] *******************************",
                                "Saturday 21 March 2026  01:07:36 +0000 (0:00:04.295)       0:00:53.402 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Get service project id] ****************************************",
                                "Saturday 21 March 2026  01:07:40 +0000 (0:00:04.471)       0:00:57.874 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Create security groups for octavia] ****************************",
                                "Saturday 21 March 2026  01:07:43 +0000 (0:00:02.994)       0:01:00.869 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=lb-mgmt-sec-grp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=lb-health-mgr-sec-grp)\u001b[0m",
                                "",
                                "TASK [octavia : Add rules for security groups] *********************************",
                                "Saturday 21 March 2026  01:07:52 +0000 (0:00:08.546)       0:01:09.415 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'icmp'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'tcp', 'src_port': 22, 'dst_port': 22}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'tcp', 'src_port': '9443', 'dst_port': '9443'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-health-mgr-sec-grp', 'enabled': True}, {'protocol': 'udp', 'src_port': '5555', 'dst_port': '5555'}])\u001b[0m",
                                "",
                                "TASK [octavia : Create loadbalancer management network] ************************",
                                "Saturday 21 March 2026  01:08:07 +0000 (0:00:15.225)       0:01:24.641 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Create loadbalancer management subnet] *************************",
                                "Saturday 21 March 2026  01:08:11 +0000 (0:00:03.946)       0:01:28.587 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Create loadbalancer management router for IPv6] ****************",
                                "Saturday 21 March 2026  01:08:15 +0000 (0:00:04.377)       0:01:32.965 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Update loadbalancer management subnet] *************************",
                                "Saturday 21 March 2026  01:08:15 +0000 (0:00:00.219)       0:01:33.185 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Saturday 21 March 2026  01:08:20 +0000 (0:00:04.931)       0:01:38.117 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/hm-interface.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [octavia : Create ports for Octavia health-manager nodes] *****************",
                                "Saturday 21 March 2026  01:08:21 +0000 (0:00:00.796)       0:01:38.913 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Update Octavia health manager port host_id] ********************",
                                "Saturday 21 March 2026  01:08:27 +0000 (0:00:05.439)       0:01:44.353 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [octavia : Add Octavia port to openvswitch br-int] ************************",
                                "Saturday 21 March 2026  01:08:31 +0000 (0:00:04.231)       0:01:48.584 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Install isc-dhcp-client package] *******************************",
                                "Saturday 21 March 2026  01:08:32 +0000 (0:00:00.723)       0:01:49.308 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Create octavia dhclient conf] **********************************",
                                "Saturday 21 March 2026  01:08:33 +0000 (0:00:01.411)       0:01:50.719 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Create octavia-interface service] ******************************",
                                "Saturday 21 March 2026  01:08:34 +0000 (0:00:01.141)       0:01:51.860 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Restart octavia-interface.service if required] *****************",
                                "Saturday 21 March 2026  01:08:35 +0000 (0:00:01.112)       0:01:52.973 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Enable and start octavia-interface.service] ********************",
                                "Saturday 21 March 2026  01:08:37 +0000 (0:00:02.052)       0:01:55.026 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Wait for interface ohm0 ip appear] *****************************",
                                "Saturday 21 March 2026  01:08:39 +0000 (0:00:01.436)       0:01:56.462 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Gather facts] **************************************************",
                                "Saturday 21 March 2026  01:08:39 +0000 (0:00:00.530)       0:01:56.992 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Saturday 21 March 2026  01:08:41 +0000 (0:00:02.275)       0:01:59.268 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/get_resources_info.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [octavia : Get amphora flavor info] ***************************************",
                                "Saturday 21 March 2026  01:08:42 +0000 (0:00:00.688)       0:01:59.957 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Get service project id] ****************************************",
                                "Saturday 21 March 2026  01:08:46 +0000 (0:00:03.444)       0:02:03.402 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Get security groups for octavia] *******************************",
                                "Saturday 21 March 2026  01:08:49 +0000 (0:00:02.931)       0:02:06.334 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=lb-mgmt-sec-grp)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=lb-health-mgr-sec-grp)\u001b[0m",
                                "",
                                "TASK [octavia : Get loadbalancer management network] ***************************",
                                "Saturday 21 March 2026  01:08:56 +0000 (0:00:07.195)       0:02:13.530 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Set octavia resources facts] ***********************************",
                                "Saturday 21 March 2026  01:08:59 +0000 (0:00:03.095)       0:02:16.625 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Ensuring config directories exist] *****************************",
                                "Saturday 21 March 2026  01:08:59 +0000 (0:00:00.333)       0:02:16.958 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [octavia : Check if policies shall be overwritten] ************************",
                                "Saturday 21 March 2026  01:09:02 +0000 (0:00:02.475)       0:02:19.434 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Set octavia policy file] ***************************************",
                                "Saturday 21 March 2026  01:09:02 +0000 (0:00:00.128)       0:02:19.563 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Copying over existing policy file] *****************************",
                                "Saturday 21 March 2026  01:09:02 +0000 (0:00:00.268)       0:02:19.831 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Saturday 21 March 2026  01:09:03 +0000 (0:00:00.663)       0:02:20.494 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : octavia | Copying over extra CA certificates] ********",
                                "Saturday 21 March 2026  01:09:03 +0000 (0:00:00.666)       0:02:21.161 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : octavia | Copying over backend internal TLS certificate] ***",
                                "Saturday 21 March 2026  01:09:08 +0000 (0:00:04.239)       0:02:25.400 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : octavia | Copying over backend internal TLS key] *****",
                                "Saturday 21 March 2026  01:09:08 +0000 (0:00:00.678)       0:02:26.079 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Copying over config.json files for services] *******************",
                                "Saturday 21 March 2026  01:09:09 +0000 (0:00:01.021)       0:02:27.101 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [octavia : Copying over octavia-wsgi.conf] ********************************",
                                "Saturday 21 March 2026  01:09:14 +0000 (0:00:05.045)       0:02:32.146 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m",
                                "",
                                "TASK [octavia : Copying over octavia.conf] *************************************",
                                "Saturday 21 March 2026  01:09:16 +0000 (0:00:01.596)       0:02:33.742 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [octavia : Copying over Octavia SSH key] **********************************",
                                "Saturday 21 March 2026  01:09:32 +0000 (0:00:16.466)       0:02:50.209 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Copying certificate files for octavia-worker] ******************",
                                "Saturday 21 March 2026  01:09:34 +0000 (0:00:02.042)       0:02:52.251 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m",
                                "",
                                "TASK [octavia : Copying certificate files for octavia-housekeeping] ************",
                                "Saturday 21 March 2026  01:09:40 +0000 (0:00:05.950)       0:02:58.202 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m",
                                "",
                                "TASK [octavia : Copying certificate files for octavia-health-manager] **********",
                                "Saturday 21 March 2026  01:09:45 +0000 (0:00:05.026)       0:03:03.229 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m",
                                "",
                                "TASK [octavia : Check octavia containers] **************************************",
                                "Saturday 21 March 2026  01:09:51 +0000 (0:00:05.076)       0:03:08.306 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Saturday 21 March 2026  01:09:54 +0000 (0:00:03.299)       0:03:11.605 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Creating Octavia database] *************************************",
                                "Saturday 21 March 2026  01:09:54 +0000 (0:00:00.455)       0:03:12.061 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Creating Octavia persistence database] *************************",
                                "Saturday 21 March 2026  01:09:56 +0000 (0:00:01.896)       0:03:13.957 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Creating Octavia database user and setting permissions] ********",
                                "Saturday 21 March 2026  01:09:58 +0000 (0:00:01.981)       0:03:15.938 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Creating Octavia persistence database user and setting permissions] ***",
                                "Saturday 21 March 2026  01:10:00 +0000 (0:00:02.202)       0:03:18.141 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Running Octavia bootstrap container] ***************************",
                                "Saturday 21 March 2026  01:10:03 +0000 (0:00:02.440)       0:03:20.582 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Flush handlers] ************************************************",
                                "Saturday 21 March 2026  01:10:24 +0000 (0:00:21.649)       0:03:42.231 ******** ",
                                "",
                                "TASK [octavia : Flush handlers] ************************************************",
                                "Saturday 21 March 2026  01:10:24 +0000 (0:00:00.069)       0:03:42.301 ******** ",
                                "",
                                "TASK [octavia : Flush handlers] ************************************************",
                                "Saturday 21 March 2026  01:10:25 +0000 (0:00:00.066)       0:03:42.368 ******** ",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-api container] **********************",
                                "Saturday 21 March 2026  01:10:25 +0000 (0:00:00.066)       0:03:42.434 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-driver-agent container] *************",
                                "Saturday 21 March 2026  01:10:35 +0000 (0:00:10.446)       0:03:52.880 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-health-manager container] ***********",
                                "Saturday 21 March 2026  01:10:45 +0000 (0:00:10.342)       0:04:03.223 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-housekeeping container] *************",
                                "Saturday 21 March 2026  01:10:55 +0000 (0:00:10.042)       0:04:13.266 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-worker container] *******************",
                                "Saturday 21 March 2026  01:11:04 +0000 (0:00:08.205)       0:04:21.471 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=57  \u001b[0m \u001b[0;33mchanged=38  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=33  \u001b[0m \u001b[0;33mchanged=22  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=33  \u001b[0m \u001b[0;33mchanged=22  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  01:11:14 +0000 (0:00:10.161)       0:04:31.633 ******** ",
                                "=============================================================================== ",
                                "octavia : Running Octavia bootstrap container -------------------------- 21.65s",
                                "octavia : Copying over octavia.conf ------------------------------------ 16.47s",
                                "octavia : Add rules for security groups -------------------------------- 15.23s",
                                "octavia : Adding octavia related roles --------------------------------- 14.92s",
                                "octavia : Restart octavia-api container -------------------------------- 10.45s",
                                "octavia : Restart octavia-driver-agent container ----------------------- 10.34s",
                                "octavia : Restart octavia-worker container ----------------------------- 10.16s",
                                "octavia : Restart octavia-health-manager container --------------------- 10.04s",
                                "octavia : Create security groups for octavia ---------------------------- 8.55s",
                                "octavia : Restart octavia-housekeeping container ------------------------ 8.21s",
                                "service-ks-register : octavia | Granting user roles --------------------- 7.99s",
                                "service-ks-register : octavia | Creating users -------------------------- 7.48s",
                                "octavia : Get security groups for octavia ------------------------------- 7.20s",
                                "service-ks-register : octavia | Creating endpoints ---------------------- 6.48s",
                                "octavia : Copying certificate files for octavia-worker ------------------ 5.95s",
                                "octavia : Create ports for Octavia health-manager nodes ----------------- 5.44s",
                                "octavia : Copying certificate files for octavia-health-manager ---------- 5.08s",
                                "octavia : Copying over config.json files for services ------------------- 5.05s",
                                "octavia : Copying certificate files for octavia-housekeeping ------------ 5.03s",
                                "octavia : Update loadbalancer management subnet ------------------------- 4.93s",
                                "\u001b[32m2026-03-21 01:11:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-21 01:11:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "",
                                "",
                                "--> DEPLOY IN A NUTSHELL -- END -- Sat Mar 21 01:12:16 UTC 2026"
                            ],
                            "zuul_log_id": "0242ac1f-6606-9900-9c98-00000000001c-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:12:16.987412Z",
                            "start": "2026-03-21T00:39:34.090522Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-00000000001c",
                        "name": "Deploy in a nutshell"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed04/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.193.136",
                                "/opt/configuration/scripts/bootstrap.sh"
                            ],
                            "delta": "0:02:58.373260",
                            "end": "2026-03-21 01:15:15.824906",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed04/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.193.136 /opt/configuration/scripts/bootstrap.sh",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-9900-9c98-00000000001d-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-21 01:12:17.451646",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "\n# BOOTSTRAP\n\n+ set -e\n+ echo\n+ echo '# BOOTSTRAP'\n+ echo\n+ sh -c /opt/configuration/scripts/bootstrap-services.sh\n+ set -e\n+ sh -c /opt/configuration/scripts/bootstrap/300-openstack.sh\n\u001b[32m2026-03-21 01:12:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task fda14370-e934-421c-b876-69a027813535 (flavor-manager) has been started and output is visible here.\u001b[0m\n\u001b[32m2026-03-21 01:12:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1L-1 created\u001b[0m\n\u001b[32m2026-03-21 01:12:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1L-1-5 created\u001b[0m\n\u001b[32m2026-03-21 01:12:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-2 created\u001b[0m\n\u001b[32m2026-03-21 01:12:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-2-5 created\u001b[0m\n\u001b[32m2026-03-21 01:12:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-4 created\u001b[0m\n\u001b[32m2026-03-21 01:12:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-4-10 created\u001b[0m\n\u001b[32m2026-03-21 01:12:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-8 created\u001b[0m\n\u001b[32m2026-03-21 01:12:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-8-20 created\u001b[0m\n\u001b[32m2026-03-21 01:12:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4 created\u001b[0m\n\u001b[32m2026-03-21 01:12:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4-10 created\u001b[0m\n\u001b[32m2026-03-21 01:12:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-8 created\u001b[0m\n\u001b[32m2026-03-21 01:12:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-8-20 created\u001b[0m\n\u001b[32m2026-03-21 01:12:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-16 created\u001b[0m\n\u001b[32m2026-03-21 01:12:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-16-50 created\u001b[0m\n\u001b[32m2026-03-21 01:12:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8 created\u001b[0m\n\u001b[32m2026-03-21 01:12:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8-20 created\u001b[0m\n\u001b[32m2026-03-21 01:12:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-16 created\u001b[0m\n\u001b[32m2026-03-21 01:12:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-16-50 created\u001b[0m\n\u001b[32m2026-03-21 01:12:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-32 created\u001b[0m\n\u001b[32m2026-03-21 01:12:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-32-100 created\u001b[0m\n\u001b[32m2026-03-21 01:12:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-16 created\u001b[0m\n\u001b[32m2026-03-21 01:12:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-16-50 created\u001b[0m\n\u001b[32m2026-03-21 01:12:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32 created\u001b[0m\n\u001b[32m2026-03-21 01:12:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32-100 created\u001b[0m\n\u001b[32m2026-03-21 01:12:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-16V-32 created\u001b[0m\n\u001b[32m2026-03-21 01:12:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-16V-32-100 created\u001b[0m\n\u001b[32m2026-03-21 01:12:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4-20s created\u001b[0m\n\u001b[32m2026-03-21 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8-50s created\u001b[0m\n\u001b[32m2026-03-21 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-16-100s created\u001b[0m\n\u001b[32m2026-03-21 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32-100s created\u001b[0m\n\u001b[32m2026-03-21 01:12:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play bootstrap-basic in environment openstack\u001b[0m\n\u001b[32m2026-03-21 01:12:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of bootstrap-basic.\u001b[0m\n\u001b[32m2026-03-21 01:12:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8dce070c-4535-47a1-9184-099bb0b7080a (bootstrap-basic) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 01:12:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 8dce070c-4535-47a1-9184-099bb0b7080a (bootstrap-basic) has been started and output is visible here.\u001b[0m\n\nPLAY [Bootstrap basic OpenStack services] **************************************\n\nTASK [Gathering Facts] *********************************************************\nSaturday 21 March 2026  01:12:45 +0000 (0:00:00.100)       0:00:00.100 ******** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nTASK [Get volume type LUKS] ****************************************************\nSaturday 21 March 2026  01:12:47 +0000 (0:00:01.933)       0:00:02.033 ******** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nTASK [Create volume type LUKS] *************************************************\nSaturday 21 March 2026  01:12:55 +0000 (0:00:08.338)       0:00:10.371 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create public network] ***************************************************\nSaturday 21 March 2026  01:13:03 +0000 (0:00:08.069)       0:00:18.441 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Set public network to default] *******************************************\nSaturday 21 March 2026  01:13:08 +0000 (0:00:04.852)       0:00:23.293 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create public subnet] ****************************************************\nSaturday 21 March 2026  01:13:14 +0000 (0:00:06.343)       0:00:29.636 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create default IPv4 subnet pool] *****************************************\nSaturday 21 March 2026  01:13:19 +0000 (0:00:04.631)       0:00:34.268 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create manager role] *****************************************************\nSaturday 21 March 2026  01:13:23 +0000 (0:00:03.856)       0:00:38.124 ******** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  01:13:27 +0000 (0:00:03.583)       0:00:41.708 ******** \n=============================================================================== \nGet volume type LUKS ---------------------------------------------------- 8.34s\nCreate volume type LUKS ------------------------------------------------- 8.07s\nSet public network to default ------------------------------------------- 6.34s\nCreate public network --------------------------------------------------- 4.85s\nCreate public subnet ---------------------------------------------------- 4.63s\nCreate default IPv4 subnet pool ----------------------------------------- 3.86s\nCreate manager role ----------------------------------------------------- 3.58s\nGathering Facts --------------------------------------------------------- 1.93s\n\u001b[32m2026-03-21 01:13:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task a7318ba2-f60b-4996-bf21-c4bcfeb025f1 (image-manager) has been started and output is visible here.\u001b[0m\n\u001b[32m2026-03-21 01:13:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.2'\u001b[0m\n\u001b[32m2026-03-21 01:13:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img: 302\u001b[0m\n\u001b[32m2026-03-21 01:13:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image Cirros 0.6.2\u001b[0m\n\u001b[32m2026-03-21 01:13:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img\u001b[0m\n\u001b[32m2026-03-21 01:13:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m\n\u001b[32m2026-03-21 01:13:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-03-21 01:13:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'Cirros 0.6.2' successfully completed, reloading images\u001b[0m\n\u001b[32m2026-03-21 01:13:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'Cirros 0.6.2'\u001b[0m\n\u001b[32m2026-03-21 01:13:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 0.6.2\u001b[0m\n\u001b[32m2026-03-21 01:13:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = cirros\u001b[0m\n\u001b[32m2026-03-21 01:13:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:cirros\u001b[0m\n\u001b[32m2026-03-21 01:13:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m\n\u001b[32m2026-03-21 01:13:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m\n\u001b[32m2026-03-21 01:13:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m\n\u001b[32m2026-03-21 01:13:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m\n\u001b[32m2026-03-21 01:13:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m\n\u001b[32m2026-03-21 01:13:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m\n\u001b[32m2026-03-21 01:13:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: cirros\u001b[0m\n\u001b[32m2026-03-21 01:13:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: minimal\u001b[0m\n\u001b[32m2026-03-21 01:13:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: never\u001b[0m\n\u001b[32m2026-03-21 01:13:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: none\u001b[0m\n\u001b[32m2026-03-21 01:13:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m\n\u001b[32m2026-03-21 01:13:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: Cirros\u001b[0m\n\u001b[32m2026-03-21 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: Cirros\u001b[0m\n\u001b[32m2026-03-21 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 0.6.2\u001b[0m\n\u001b[32m2026-03-21 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: cirros\u001b[0m\n\u001b[32m2026-03-21 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 0.6.2\u001b[0m\n\u001b[32m2026-03-21 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img\u001b[0m\n\u001b[32m2026-03-21 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2023-05-30\u001b[0m\n\u001b[32m2026-03-21 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'Cirros 0.6.2'\u001b[0m\n\u001b[32m2026-03-21 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'Cirros 0.6.2'\u001b[0m\n\u001b[32m2026-03-21 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting visibility of 'Cirros 0.6.2' to 'public'\u001b[0m\n\u001b[32m2026-03-21 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.3'\u001b[0m\n\u001b[32m2026-03-21 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img: 302\u001b[0m\n\u001b[32m2026-03-21 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image Cirros 0.6.3\u001b[0m\n\u001b[32m2026-03-21 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img\u001b[0m\n\u001b[32m2026-03-21 01:13:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m\n\u001b[32m2026-03-21 01:13:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-03-21 01:14:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'Cirros 0.6.3' successfully completed, reloading images\u001b[0m\n\u001b[32m2026-03-21 01:14:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'Cirros 0.6.3'\u001b[0m\n\u001b[32m2026-03-21 01:14:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 0.6.3\u001b[0m\n\u001b[32m2026-03-21 01:14:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = cirros\u001b[0m\n\u001b[32m2026-03-21 01:14:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:cirros\u001b[0m\n\u001b[32m2026-03-21 01:14:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m\n\u001b[32m2026-03-21 01:14:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m\n\u001b[32m2026-03-21 01:14:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m\n\u001b[32m2026-03-21 01:14:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m\n\u001b[32m2026-03-21 01:14:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m\n\u001b[32m2026-03-21 01:14:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m\n\u001b[32m2026-03-21 01:14:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: cirros\u001b[0m\n\u001b[32m2026-03-21 01:14:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: minimal\u001b[0m\n\u001b[32m2026-03-21 01:14:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: never\u001b[0m\n\u001b[32m2026-03-21 01:14:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: none\u001b[0m\n\u001b[32m2026-03-21 01:14:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m\n\u001b[32m2026-03-21 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: Cirros\u001b[0m\n\u001b[32m2026-03-21 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: Cirros\u001b[0m\n\u001b[32m2026-03-21 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 0.6.3\u001b[0m\n\u001b[32m2026-03-21 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: cirros\u001b[0m\n\u001b[32m2026-03-21 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 0.6.3\u001b[0m\n\u001b[32m2026-03-21 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img\u001b[0m\n\u001b[32m2026-03-21 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2024-09-26\u001b[0m\n\u001b[32m2026-03-21 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'Cirros 0.6.3'\u001b[0m\n\u001b[32m2026-03-21 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'Cirros 0.6.3'\u001b[0m\n\u001b[32m2026-03-21 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting visibility of 'Cirros 0.6.3' to 'public'\u001b[0m\n+ sh -c /opt/configuration/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh\n\u001b[32m2026-03-21 01:14:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mdate: 2026-03-20\u001b[0m\n\u001b[32m2026-03-21 01:14:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mimage: octavia-amphora-haproxy-2024.2.20260320.qcow2\u001b[0m\n\u001b[32m2026-03-21 01:14:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1murl: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260320.qcow2\u001b[0m\n\u001b[32m2026-03-21 01:14:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mchecksum_url: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260320.qcow2.CHECKSUM\u001b[0m\n\u001b[32m2026-03-21 01:14:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mchecksum: 71110bf690c634ec250d031ee3a7f0333fd8b734d34f22a3f5110f25facfb804\u001b[0m\n\u001b[32m2026-03-21 01:14:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 24d33f5e-0b65-4287-b5c3-4f840738a85c (image-manager) has been started and output is visible here.\u001b[0m\n\u001b[32m2026-03-21 01:14:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'OpenStack Octavia Amphora 2026-03-20'\u001b[0m\n\u001b[32m2026-03-21 01:14:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260320.qcow2: 200\u001b[0m\n\u001b[32m2026-03-21 01:14:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image OpenStack Octavia Amphora 2026-03-20\u001b[0m\n\u001b[32m2026-03-21 01:14:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260320.qcow2\u001b[0m\n\u001b[32m2026-03-21 01:14:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m\n\u001b[32m2026-03-21 01:14:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-03-21 01:14:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-03-21 01:14:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-03-21 01:14:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-03-21 01:15:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-03-21 01:15:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'OpenStack Octavia Amphora 2026-03-20' successfully completed, reloading images\u001b[0m\n\u001b[32m2026-03-21 01:15:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'OpenStack Octavia Amphora 2026-03-20'\u001b[0m\n\u001b[32m2026-03-21 01:15:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 2026-03-20\u001b[0m\n\u001b[32m2026-03-21 01:15:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = ubuntu\u001b[0m\n\u001b[32m2026-03-21 01:15:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag amphora\u001b[0m\n\u001b[32m2026-03-21 01:15:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:ubuntu\u001b[0m\n\u001b[32m2026-03-21 01:15:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m\n\u001b[32m2026-03-21 01:15:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m\n\u001b[32m2026-03-21 01:15:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m\n\u001b[32m2026-03-21 01:15:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m\n\u001b[32m2026-03-21 01:15:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m\n\u001b[32m2026-03-21 01:15:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m\n\u001b[32m2026-03-21 01:15:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: ubuntu\u001b[0m\n\u001b[32m2026-03-21 01:15:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: quarterly\u001b[0m\n\u001b[32m2026-03-21 01:15:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: last-1\u001b[0m\n\u001b[32m2026-03-21 01:15:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m\n\u001b[32m2026-03-21 01:15:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: network\u001b[0m\n\u001b[32m2026-03-21 01:15:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: OpenStack Octavia Amphora\u001b[0m\n\u001b[32m2026-03-21 01:15:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: OpenStack Octavia Amphora\u001b[0m\n\u001b[32m2026-03-21 01:15:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 2026-03-20\u001b[0m\n\u001b[32m2026-03-21 01:15:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: ubuntu\u001b[0m\n\u001b[32m2026-03-21 01:15:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 2026-03-20\u001b[0m\n\u001b[32m2026-03-21 01:15:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260320.qcow2\u001b[0m\n\u001b[32m2026-03-21 01:15:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2026-03-20\u001b[0m\n\u001b[32m2026-03-21 01:15:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'OpenStack Octavia Amphora 2026-03-20'\u001b[0m\n\u001b[32m2026-03-21 01:15:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'OpenStack Octavia Amphora 2026-03-20'\u001b[0m\n\u001b[32m2026-03-21 01:15:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.3' (removal candidate)\u001b[0m\n\u001b[32m2026-03-21 01:15:15\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[33m\u001b[1mNo image definition found for 'Cirros 0.6.3', image will be ignored\u001b[0m\n\u001b[32m2026-03-21 01:15:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.2' (removal candidate)\u001b[0m\n\u001b[32m2026-03-21 01:15:15\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[33m\u001b[1mNo image definition found for 'Cirros 0.6.2', image will be ignored\u001b[0m",
                            "stdout_lines": [
                                "",
                                "# BOOTSTRAP",
                                "",
                                "+ set -e",
                                "+ echo",
                                "+ echo '# BOOTSTRAP'",
                                "+ echo",
                                "+ sh -c /opt/configuration/scripts/bootstrap-services.sh",
                                "+ set -e",
                                "+ sh -c /opt/configuration/scripts/bootstrap/300-openstack.sh",
                                "\u001b[32m2026-03-21 01:12:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task fda14370-e934-421c-b876-69a027813535 (flavor-manager) has been started and output is visible here.\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1L-1 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1L-1-5 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-2 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-2-5 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-4 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-4-10 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-8 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-8-20 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4-10 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-8 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-8-20 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-16 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-16-50 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8-20 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-16 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-16-50 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-32 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-32-100 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-16 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-16-50 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32-100 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-16V-32 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-16V-32-100 created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4-20s created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8-50s created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-16-100s created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32-100s created\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play bootstrap-basic in environment openstack\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of bootstrap-basic.\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8dce070c-4535-47a1-9184-099bb0b7080a (bootstrap-basic) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 01:12:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 8dce070c-4535-47a1-9184-099bb0b7080a (bootstrap-basic) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Bootstrap basic OpenStack services] **************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Saturday 21 March 2026  01:12:45 +0000 (0:00:00.100)       0:00:00.100 ******** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "TASK [Get volume type LUKS] ****************************************************",
                                "Saturday 21 March 2026  01:12:47 +0000 (0:00:01.933)       0:00:02.033 ******** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "TASK [Create volume type LUKS] *************************************************",
                                "Saturday 21 March 2026  01:12:55 +0000 (0:00:08.338)       0:00:10.371 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create public network] ***************************************************",
                                "Saturday 21 March 2026  01:13:03 +0000 (0:00:08.069)       0:00:18.441 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Set public network to default] *******************************************",
                                "Saturday 21 March 2026  01:13:08 +0000 (0:00:04.852)       0:00:23.293 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create public subnet] ****************************************************",
                                "Saturday 21 March 2026  01:13:14 +0000 (0:00:06.343)       0:00:29.636 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create default IPv4 subnet pool] *****************************************",
                                "Saturday 21 March 2026  01:13:19 +0000 (0:00:04.631)       0:00:34.268 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create manager role] *****************************************************",
                                "Saturday 21 March 2026  01:13:23 +0000 (0:00:03.856)       0:00:38.124 ******** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  01:13:27 +0000 (0:00:03.583)       0:00:41.708 ******** ",
                                "=============================================================================== ",
                                "Get volume type LUKS ---------------------------------------------------- 8.34s",
                                "Create volume type LUKS ------------------------------------------------- 8.07s",
                                "Set public network to default ------------------------------------------- 6.34s",
                                "Create public network --------------------------------------------------- 4.85s",
                                "Create public subnet ---------------------------------------------------- 4.63s",
                                "Create default IPv4 subnet pool ----------------------------------------- 3.86s",
                                "Create manager role ----------------------------------------------------- 3.58s",
                                "Gathering Facts --------------------------------------------------------- 1.93s",
                                "\u001b[32m2026-03-21 01:13:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task a7318ba2-f60b-4996-bf21-c4bcfeb025f1 (image-manager) has been started and output is visible here.\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.2'\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img: 302\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image Cirros 0.6.2\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'Cirros 0.6.2' successfully completed, reloading images\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'Cirros 0.6.2'\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 0.6.2\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = cirros\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:cirros\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: cirros\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: minimal\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: never\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: none\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: Cirros\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: Cirros\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 0.6.2\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: cirros\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 0.6.2\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2023-05-30\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'Cirros 0.6.2'\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'Cirros 0.6.2'\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting visibility of 'Cirros 0.6.2' to 'public'\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.3'\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img: 302\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image Cirros 0.6.3\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m",
                                "\u001b[32m2026-03-21 01:13:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'Cirros 0.6.3' successfully completed, reloading images\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'Cirros 0.6.3'\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 0.6.3\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = cirros\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:cirros\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: cirros\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: minimal\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: never\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: none\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: Cirros\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: Cirros\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 0.6.3\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: cirros\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 0.6.3\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2024-09-26\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'Cirros 0.6.3'\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'Cirros 0.6.3'\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting visibility of 'Cirros 0.6.3' to 'public'\u001b[0m",
                                "+ sh -c /opt/configuration/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh",
                                "\u001b[32m2026-03-21 01:14:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mdate: 2026-03-20\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mimage: octavia-amphora-haproxy-2024.2.20260320.qcow2\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1murl: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260320.qcow2\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mchecksum_url: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260320.qcow2.CHECKSUM\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mchecksum: 71110bf690c634ec250d031ee3a7f0333fd8b734d34f22a3f5110f25facfb804\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 24d33f5e-0b65-4287-b5c3-4f840738a85c (image-manager) has been started and output is visible here.\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'OpenStack Octavia Amphora 2026-03-20'\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260320.qcow2: 200\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image OpenStack Octavia Amphora 2026-03-20\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260320.qcow2\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-03-21 01:14:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'OpenStack Octavia Amphora 2026-03-20' successfully completed, reloading images\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'OpenStack Octavia Amphora 2026-03-20'\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 2026-03-20\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = ubuntu\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag amphora\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:ubuntu\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: ubuntu\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: quarterly\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: last-1\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: network\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: OpenStack Octavia Amphora\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: OpenStack Octavia Amphora\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 2026-03-20\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: ubuntu\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 2026-03-20\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260320.qcow2\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2026-03-20\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'OpenStack Octavia Amphora 2026-03-20'\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'OpenStack Octavia Amphora 2026-03-20'\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.3' (removal candidate)\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:15\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[33m\u001b[1mNo image definition found for 'Cirros 0.6.3', image will be ignored\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.2' (removal candidate)\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:15\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[33m\u001b[1mNo image definition found for 'Cirros 0.6.2', image will be ignored\u001b[0m"
                            ],
                            "zuul_log_id": "0242ac1f-6606-9900-9c98-00000000001d-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:15:16.250323Z",
                            "start": "2026-03-21T01:12:17.140416Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-00000000001d",
                        "name": "Bootstrap services"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed04/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.193.136",
                                "/opt/configuration/scripts/check.sh"
                            ],
                            "delta": "0:18:21.821190",
                            "end": "2026-03-21 01:33:38.393082",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed04/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.193.136 /opt/configuration/scripts/check.sh",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-9900-9c98-00000000001e-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-21 01:15:16.571892",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/configuration/scripts/manager-version.sh\n+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml\n\n# CHECK\n\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n+ echo\n+ echo '# CHECK'\n+ echo\n+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2\n++ semver latest 5.0.0\n\n## Containers @ testbed-manager\n\n+ [[ -1 -eq -1 ]]\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ echo\n+ echo '## Containers @ testbed-manager'\n+ echo\n+ osism container testbed-manager ps\n\u001b[32m2026-03-21 01:15:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCreating empty known_hosts file: /share/known_hosts\u001b[0m\nCONTAINER ID   IMAGE                                                           COMMAND                  CREATED          STATUS                    PORTS                                                                             NAMES\n79dcf2ba75f6   registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2   \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes                                                                                               prometheus_blackbox_exporter\nd0253e1bc915   registry.osism.tech/kolla/prometheus-alertmanager:2024.2        \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                                                                                               prometheus_alertmanager\ndebdfb3eb46d   registry.osism.tech/kolla/prometheus-cadvisor:2024.2            \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                                                                                               prometheus_cadvisor\nae733499c9e4   registry.osism.tech/kolla/prometheus-node-exporter:2024.2       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                                                                                               prometheus_node_exporter\nd34ecdad237a   registry.osism.tech/kolla/prometheus-v2-server:2024.2           \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                                                                                               prometheus_server\n32ffe7d6598e   registry.osism.tech/osism/cephclient:reef                       \"/usr/bin/dumb-init \u2026\"   16 minutes ago   Up 15 minutes                                                                                               cephclient\nb11564e80c95   registry.osism.tech/kolla/cron:2024.2                           \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes                                                                                               cron\nf36f602ac7cc   registry.osism.tech/kolla/kolla-toolbox:2024.2                  \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes                                                                                               kolla_toolbox\n87666c63b6d1   registry.osism.tech/kolla/fluentd:2024.2                        \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                                                                                               fluentd\n3d86d9380bd4   phpmyadmin/phpmyadmin:5.2                                       \"/docker-entrypoint.\u2026\"   29 minutes ago   Up 29 minutes (healthy)   80/tcp                                                                            phpmyadmin\nd2d538d7dc83   registry.osism.tech/osism/openstackclient:2024.2                \"/usr/bin/dumb-init \u2026\"   30 minutes ago   Up 29 minutes                                                                                               openstackclient\n2cd72298f354   registry.osism.tech/osism/homer:v25.10.1                        \"/bin/sh /entrypoint\u2026\"   30 minutes ago   Up 29 minutes (healthy)   8080/tcp                                                                          homer\nf5125536ac8a   registry.osism.tech/dockerhub/ubuntu/squid:6.1-23.10_beta       \"entrypoint.sh -f /e\u2026\"   52 minutes ago   Up 52 minutes (healthy)   192.168.16.5:3128->3128/tcp                                                       squid\n8031abebbadc   registry.osism.tech/osism/inventory-reconciler:latest           \"/sbin/tini -- /entr\u2026\"   57 minutes ago   Up 36 minutes (healthy)                                                                                     manager-inventory_reconciler-1\n98e5144f44c7   registry.osism.tech/osism/osism-ansible:latest                  \"/entrypoint.sh osis\u2026\"   57 minutes ago   Up 36 minutes (healthy)                                                                                     osism-ansible\n198ac8812cab   registry.osism.tech/osism/kolla-ansible:2024.2                  \"/entrypoint.sh osis\u2026\"   57 minutes ago   Up 36 minutes (healthy)                                                                                     kolla-ansible\nddda498069fb   registry.osism.tech/osism/osism-kubernetes:latest               \"/entrypoint.sh osis\u2026\"   57 minutes ago   Up 36 minutes (healthy)                                                                                     osism-kubernetes\n2a2e2ce2531b   registry.osism.tech/osism/ceph-ansible:reef                     \"/entrypoint.sh osis\u2026\"   57 minutes ago   Up 36 minutes (healthy)                                                                                     ceph-ansible\nc3b8f043a57a   registry.osism.tech/osism/ara-server:1.7.3                      \"sh -c '/wait && /ru\u2026\"   57 minutes ago   Up 36 minutes (healthy)   8000/tcp                                                                          manager-ara-server-1\n36eb7f15777c   registry.osism.tech/osism/osism-frontend:latest                 \"docker-entrypoint.s\u2026\"   57 minutes ago   Up 37 minutes             192.168.16.5:3000->3000/tcp                                                       osism-frontend\n2cfcad7bb404   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   57 minutes ago   Up 37 minutes (healthy)                                                                                     manager-openstack-1\n538cf49a52a2   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   57 minutes ago   Up 37 minutes (healthy)                                                                                     manager-listener-1\n8cc90e9c6172   registry.osism.tech/dockerhub/library/redis:7.4.7-alpine        \"docker-entrypoint.s\u2026\"   57 minutes ago   Up 37 minutes (healthy)   6379/tcp                                                                          manager-redis-1\n72d89da16cba   registry.osism.tech/dockerhub/library/mariadb:11.8.4            \"docker-entrypoint.s\u2026\"   57 minutes ago   Up 37 minutes (healthy)   3306/tcp                                                                          manager-mariadb-1\n2d9151476f4b   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   57 minutes ago   Up 37 minutes (healthy)                                                                                     manager-beat-1\n2f67b711771f   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- sleep\u2026\"   57 minutes ago   Up 37 minutes (healthy)                                                                                     osismclient\nc309a338448f   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   57 minutes ago   Up 37 minutes (healthy)                                                                                     manager-flower-1\nbc21652d95e6   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   57 minutes ago   Up 37 minutes (healthy)   192.168.16.5:8000->8000/tcp                                                       manager-api-1\n12dae91709a5   registry.osism.tech/dockerhub/library/traefik:v3.5.0            \"/entrypoint.sh trae\u2026\"   58 minutes ago   Up 58 minutes (healthy)   192.168.16.5:80->80/tcp, 192.168.16.5:443->443/tcp, 192.168.16.5:8122->8080/tcp   traefik\n\n## Images @ testbed-manager\n\n+ echo\n+ echo '## Images @ testbed-manager'\n+ echo\n+ osism container testbed-manager images\nREPOSITORY                                               TAG              IMAGE ID       CREATED             SIZE\nregistry.osism.tech/osism/osism-ansible                  latest           5c181d8e29f7   About an hour ago   634MB\nregistry.osism.tech/osism/osism-kubernetes               latest           eb3836cac167   About an hour ago   1.24GB\nregistry.osism.tech/osism/kolla-ansible                  2024.2           6a7c8613a511   About an hour ago   631MB\nregistry.osism.tech/osism/ceph-ansible                   reef             ff5b1d600058   About an hour ago   581MB\nregistry.osism.tech/osism/osism                          latest           4102582d9e9d   About an hour ago   408MB\nregistry.osism.tech/osism/osism-frontend                 latest           fd751f4a01e0   About an hour ago   232MB\nregistry.osism.tech/osism/inventory-reconciler           latest           056df03054cc   About an hour ago   357MB\nregistry.osism.tech/osism/openstackclient                2024.2           87a2e78155c3   21 hours ago        239MB\nregistry.osism.tech/osism/cephclient                     reef             102963fef6c9   21 hours ago        453MB\nregistry.osism.tech/kolla/kolla-toolbox                  2024.2           2ff405629e62   23 hours ago        677MB\nregistry.osism.tech/kolla/cron                           2024.2           23ec8544c793   23 hours ago        275MB\nregistry.osism.tech/kolla/fluentd                        2024.2           851a383a0141   23 hours ago        589MB\nregistry.osism.tech/kolla/prometheus-alertmanager        2024.2           eb2991a15b6c   24 hours ago        413MB\nregistry.osism.tech/kolla/prometheus-node-exporter       2024.2           e9fbd7018fbf   24 hours ago        315MB\nregistry.osism.tech/kolla/prometheus-cadvisor            2024.2           1520e18eb877   24 hours ago        367MB\nregistry.osism.tech/kolla/prometheus-blackbox-exporter   2024.2           f23bcd4578b3   24 hours ago        317MB\nregistry.osism.tech/kolla/prometheus-v2-server           2024.2           ecfb9f783b6e   24 hours ago        849MB\nregistry.osism.tech/dockerhub/library/redis              7.4.7-alpine     e08bd8d5a677   7 weeks ago         41.4MB\nregistry.osism.tech/osism/homer                          v25.10.1         ea34b371c716   3 months ago        11.5MB\nregistry.osism.tech/dockerhub/library/mariadb            11.8.4           70745dd8f1d0   4 months ago        334MB\nphpmyadmin/phpmyadmin                                    5.2              e66b1f5a8c58   5 months ago        742MB\nregistry.osism.tech/osism/ara-server                     1.7.3            d1b687333f2f   6 months ago        275MB\nregistry.osism.tech/dockerhub/library/traefik            v3.5.0           11cc59587f6a   8 months ago        226MB\nregistry.osism.tech/dockerhub/ubuntu/squid               6.1-23.10_beta   34b6bbbcf74b   21 months ago       146MB\n+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2\n++ semver latest 5.0.0\n\n## Containers @ testbed-node-0\n\n+ [[ -1 -eq -1 ]]\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ echo\n+ echo '## Containers @ testbed-node-0'\n+ echo\n+ osism container testbed-node-0 ps\nCONTAINER ID   IMAGE                                                                COMMAND                  CREATED          STATUS                    PORTS     NAMES\n9d8922e858b5   registry.osism.tech/kolla/octavia-worker:2024.2                      \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_worker\na6ee23841317   registry.osism.tech/kolla/octavia-housekeeping:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_housekeeping\n4a5db920cf4d   registry.osism.tech/kolla/octavia-health-manager:2024.2              \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_health_manager\n8afd7c71e10e   registry.osism.tech/kolla/octavia-driver-agent:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes                        octavia_driver_agent\nd75168b985b3   registry.osism.tech/kolla/octavia-api:2024.2                         \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_api\n112fca436d51   registry.osism.tech/kolla/nova-novncproxy:2024.2                     \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_novncproxy\n2debb8ac5833   registry.osism.tech/kolla/nova-conductor:2024.2                      \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_conductor\n17b1c2bc2a93   registry.osism.tech/kolla/nova-api:2024.2                            \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_api\nd03f58f2a91a   registry.osism.tech/kolla/nova-scheduler:2024.2                      \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_scheduler\n079169f0167c   registry.osism.tech/kolla/grafana:2024.2                             \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes                        grafana\n9563220c0f32   registry.osism.tech/kolla/cinder-backup:2024.2                       \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_backup\n01f462bda636   registry.osism.tech/kolla/cinder-volume:2024.2                       \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_volume\n7167a3f3c3b3   registry.osism.tech/kolla/glance-api:2024.2                          \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             glance_api\n670c1d20efa4   registry.osism.tech/kolla/cinder-scheduler:2024.2                    \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_scheduler\n789c5c06779c   registry.osism.tech/kolla/cinder-api:2024.2                          \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_api\n819bdc497753   registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2   \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_elasticsearch_exporter\nba617ee789c0   registry.osism.tech/kolla/prometheus-cadvisor:2024.2                 \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_cadvisor\n9eab234cd872   registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_memcached_exporter\n1c50f7207750   registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2          \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_mysqld_exporter\nd081c69342ac   registry.osism.tech/kolla/prometheus-node-exporter:2024.2            \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_node_exporter\n60ef4c825ac7   registry.osism.tech/kolla/magnum-conductor:2024.2                    \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             magnum_conductor\n0b1d6881c3b6   registry.osism.tech/kolla/magnum-api:2024.2                          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             magnum_api\n2a8d8d114469   registry.osism.tech/kolla/neutron-server:2024.2                      \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             neutron_server\n0889cecb0ed2   registry.osism.tech/kolla/designate-worker:2024.2                    \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             designate_worker\n23631501f2a2   registry.osism.tech/kolla/placement-api:2024.2                       \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             placement_api\n85a3b0d8dd80   registry.osism.tech/kolla/designate-mdns:2024.2                      \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_mdns\n2facf6912789   registry.osism.tech/kolla/designate-producer:2024.2                  \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_producer\n9690c76363a7   registry.osism.tech/kolla/designate-central:2024.2                   \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_central\n6440e8936ce4   registry.osism.tech/kolla/designate-api:2024.2                       \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_api\n5d02b1d19aec   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mgr -\u2026\"   14 minutes ago   Up 14 minutes                       ceph-mgr-testbed-node-0\n8ccd851e0dfe   registry.osism.tech/kolla/barbican-worker:2024.2                     \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             barbican_worker\n111d3ae52613   registry.osism.tech/kolla/designate-backend-bind9:2024.2             \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_backend_bind9\n102484264d0e   registry.osism.tech/kolla/barbican-keystone-listener:2024.2          \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             barbican_keystone_listener\ne10dd320f39d   registry.osism.tech/kolla/barbican-api:2024.2                        \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             barbican_api\n3425e4fe5ea9   registry.osism.tech/kolla/keystone:2024.2                            \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             keystone\n19fdb3f3eae3   registry.osism.tech/kolla/keystone-fernet:2024.2                     \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             keystone_fernet\nabaa7f710593   registry.osism.tech/kolla/keystone-ssh:2024.2                        \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_ssh\ne75cf675d062   registry.osism.tech/kolla/horizon:2024.2                             \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             horizon\n70eba282b12c   registry.osism.tech/kolla/mariadb-server:2024.2                      \"dumb-init -- kolla_\u2026\"   19 minutes ago   Up 19 minutes (healthy)             mariadb\n38749466ddbb   registry.osism.tech/kolla/opensearch-dashboards:2024.2               \"dumb-init --single-\u2026\"   20 minutes ago   Up 20 minutes (healthy)             opensearch_dashboards\n4b878a379f89   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-crash\"    21 minutes ago   Up 21 minutes                       ceph-crash-testbed-node-0\n06727c8ef341   registry.osism.tech/kolla/opensearch:2024.2                          \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             opensearch\ne6cd19bfe23a   registry.osism.tech/kolla/keepalived:2024.2                          \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes                       keepalived\nbd8a3271595a   registry.osism.tech/kolla/proxysql:2024.2                            \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes (healthy)             proxysql\nd3d6d8a0896b   registry.osism.tech/kolla/haproxy:2024.2                             \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes (healthy)             haproxy\n60aef232a911   registry.osism.tech/kolla/ovn-northd:2024.2                          \"dumb-init --single-\u2026\"   25 minutes ago   Up 25 minutes                       ovn_northd\nc5e129f45a17   registry.osism.tech/kolla/ovn-sb-db-server:2024.2                    \"dumb-init --single-\u2026\"   25 minutes ago   Up 25 minutes                       ovn_sb_db\nebdab83132e9   registry.osism.tech/kolla/ovn-nb-db-server:2024.2                    \"dumb-init --single-\u2026\"   25 minutes ago   Up 25 minutes                       ovn_nb_db\ne75f5e68cdc6   registry.osism.tech/kolla/ovn-controller:2024.2                      \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       ovn_controller\n22929cdeac3a   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mon -\u2026\"   26 minutes ago   Up 26 minutes                       ceph-mon-testbed-node-0\n9f9c1ac5d698   registry.osism.tech/kolla/rabbitmq:2024.2                            \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             rabbitmq\n50d0a4f1eb1d   registry.osism.tech/kolla/openvswitch-vswitchd:2024.2                \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             openvswitch_vswitchd\nc27b35eea4c1   registry.osism.tech/kolla/openvswitch-db-server:2024.2               \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             openvswitch_db\nd7ac3868d9d8   registry.osism.tech/kolla/redis-sentinel:2024.2                      \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             redis_sentinel\n30b4cc84749b   registry.osism.tech/kolla/redis:2024.2                               \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             redis\n759d15301192   registry.osism.tech/kolla/memcached:2024.2                           \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             memcached\n1ccc23d1ce4f   registry.osism.tech/kolla/cron:2024.2                                \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes                       cron\nd32c9516f38b   registry.osism.tech/kolla/kolla-toolbox:2024.2                       \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                       kolla_toolbox\nde86538f2392   registry.osism.tech/kolla/fluentd:2024.2                             \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                       fluentd\n\n## Images @ testbed-node-0\n\n+ echo\n+ echo '## Images @ testbed-node-0'\n+ echo\n+ osism container testbed-node-0 images\nREPOSITORY                                                    TAG       IMAGE ID       CREATED        SIZE\nregistry.osism.tech/osism/ceph-daemon                         reef      60d930974fb1   22 hours ago   1.27GB\nregistry.osism.tech/kolla/keepalived                          2024.2    b2b59a47e345   23 hours ago   286MB\nregistry.osism.tech/kolla/grafana                             2024.2    25f6c06d0021   23 hours ago   1.04GB\nregistry.osism.tech/kolla/kolla-toolbox                       2024.2    2ff405629e62   23 hours ago   677MB\nregistry.osism.tech/kolla/cron                                2024.2    23ec8544c793   23 hours ago   275MB\nregistry.osism.tech/kolla/opensearch-dashboards               2024.2    687dcde2bcb8   23 hours ago   1.54GB\nregistry.osism.tech/kolla/opensearch                          2024.2    b409ad33e488   23 hours ago   1.57GB\nregistry.osism.tech/kolla/memcached                           2024.2    ed2bff2a6973   23 hours ago   276MB\nregistry.osism.tech/kolla/proxysql                            2024.2    b4d6db8fd341   23 hours ago   426MB\nregistry.osism.tech/kolla/fluentd                             2024.2    851a383a0141   23 hours ago   589MB\nregistry.osism.tech/kolla/rabbitmq                            2024.2    a4ab33fd70fd   23 hours ago   332MB\nregistry.osism.tech/kolla/haproxy                             2024.2    3fa2a10107a8   23 hours ago   284MB\nregistry.osism.tech/kolla/redis-sentinel                      2024.2    5f4f04f4383c   24 hours ago   282MB\nregistry.osism.tech/kolla/redis                               2024.2    533c487568ee   24 hours ago   282MB\nregistry.osism.tech/kolla/mariadb-server                      2024.2    297f8bb339a1   24 hours ago   462MB\nregistry.osism.tech/kolla/prometheus-memcached-exporter       2024.2    dd336c43b779   24 hours ago   308MB\nregistry.osism.tech/kolla/prometheus-node-exporter            2024.2    e9fbd7018fbf   24 hours ago   315MB\nregistry.osism.tech/kolla/prometheus-mysqld-exporter          2024.2    548293489d76   24 hours ago   310MB\nregistry.osism.tech/kolla/prometheus-cadvisor                 2024.2    1520e18eb877   24 hours ago   367MB\nregistry.osism.tech/kolla/prometheus-elasticsearch-exporter   2024.2    5b796764074a   24 hours ago   301MB\nregistry.osism.tech/kolla/horizon                             2024.2    50eeb61edd36   24 hours ago   1.16GB\nregistry.osism.tech/kolla/openvswitch-vswitchd                2024.2    dbcb0c63d3a2   24 hours ago   288MB\nregistry.osism.tech/kolla/openvswitch-db-server               2024.2    a8f2511eff1f   24 hours ago   288MB\nregistry.osism.tech/kolla/neutron-server                      2024.2    7c6811afdb1b   24 hours ago   1.17GB\nregistry.osism.tech/kolla/octavia-api                         2024.2    eaf97bcdf081   24 hours ago   1.06GB\nregistry.osism.tech/kolla/octavia-driver-agent                2024.2    4c4abbcba868   24 hours ago   1.06GB\nregistry.osism.tech/kolla/octavia-worker                      2024.2    ae8999de916c   24 hours ago   1.04GB\nregistry.osism.tech/kolla/octavia-housekeeping                2024.2    e15b03944567   24 hours ago   1.04GB\nregistry.osism.tech/kolla/octavia-health-manager              2024.2    41dcecaeaf89   24 hours ago   1.04GB\nregistry.osism.tech/kolla/skyline-apiserver                   2024.2    9f9c69d19c31   24 hours ago   999MB\nregistry.osism.tech/kolla/skyline-console                     2024.2    f5b7b069c854   24 hours ago   1.06GB\nregistry.osism.tech/kolla/keystone                            2024.2    a6b20f41325d   24 hours ago   1.08GB\nregistry.osism.tech/kolla/keystone-fernet                     2024.2    15aaf5c48263   24 hours ago   1.05GB\nregistry.osism.tech/kolla/keystone-ssh                        2024.2    2634cde67cfd   24 hours ago   1.05GB\nregistry.osism.tech/kolla/cinder-backup                       2024.2    5f8e22fdae27   24 hours ago   1.42GB\nregistry.osism.tech/kolla/cinder-volume                       2024.2    29801e980d03   24 hours ago   1.73GB\nregistry.osism.tech/kolla/cinder-scheduler                    2024.2    7d98f6182a5d   24 hours ago   1.41GB\nregistry.osism.tech/kolla/cinder-api                          2024.2    c16461b7e25c   24 hours ago   1.41GB\nregistry.osism.tech/kolla/barbican-worker                     2024.2    cbe9a09e642e   24 hours ago   1GB\nregistry.osism.tech/kolla/barbican-keystone-listener          2024.2    8b79c739c5aa   24 hours ago   1GB\nregistry.osism.tech/kolla/barbican-api                        2024.2    ec75be03f45f   24 hours ago   1e+03MB\nregistry.osism.tech/kolla/nova-scheduler                      2024.2    d13b758badcb   24 hours ago   1.22GB\nregistry.osism.tech/kolla/nova-api                            2024.2    fa74d9e1d80c   24 hours ago   1.22GB\nregistry.osism.tech/kolla/nova-novncproxy                     2024.2    b0f517fc2110   24 hours ago   1.38GB\nregistry.osism.tech/kolla/nova-conductor                      2024.2    cc5099f460a9   24 hours ago   1.22GB\nregistry.osism.tech/kolla/ceilometer-notification             2024.2    447d352c2733   24 hours ago   985MB\nregistry.osism.tech/kolla/ceilometer-central                  2024.2    494e092cff9e   24 hours ago   986MB\nregistry.osism.tech/kolla/glance-api                          2024.2    0c1d9d7160de   24 hours ago   1.11GB\nregistry.osism.tech/kolla/magnum-conductor                    2024.2    e4d602a3ad04   24 hours ago   1.25GB\nregistry.osism.tech/kolla/magnum-api                          2024.2    083abdbaa2fa   24 hours ago   1.14GB\nregistry.osism.tech/kolla/aodh-evaluator                      2024.2    e38bf12d2f87   24 hours ago   984MB\nregistry.osism.tech/kolla/aodh-notifier                       2024.2    d90172096345   24 hours ago   984MB\nregistry.osism.tech/kolla/aodh-listener                       2024.2    9cd4a42b1c66   24 hours ago   984MB\nregistry.osism.tech/kolla/aodh-api                            2024.2    ca349612b64e   24 hours ago   983MB\nregistry.osism.tech/kolla/placement-api                       2024.2    dfc171655fae   24 hours ago   985MB\nregistry.osism.tech/kolla/designate-worker                    2024.2    e9363cf6baa5   24 hours ago   998MB\nregistry.osism.tech/kolla/designate-producer                  2024.2    1d3f00ae0cb0   24 hours ago   994MB\nregistry.osism.tech/kolla/designate-mdns                      2024.2    b74dea85ff31   24 hours ago   994MB\nregistry.osism.tech/kolla/designate-backend-bind9             2024.2    71020f79d0b8   24 hours ago   998MB\nregistry.osism.tech/kolla/designate-central                   2024.2    e270fdf55dd9   24 hours ago   993MB\nregistry.osism.tech/kolla/designate-api                       2024.2    2b71cfe02281   24 hours ago   994MB\nregistry.osism.tech/kolla/ovn-nb-db-server                    2024.2    91492a927f10   24 hours ago   850MB\nregistry.osism.tech/kolla/ovn-sb-db-server                    2024.2    c3d787fcf696   24 hours ago   850MB\nregistry.osism.tech/kolla/ovn-controller                      2024.2    df4830242254   24 hours ago   850MB\nregistry.osism.tech/kolla/ovn-northd                          2024.2    521cb38ac60e   24 hours ago   850MB\n+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2\n++ semver latest 5.0.0\n\n## Containers @ testbed-node-1\n\n+ [[ -1 -eq -1 ]]\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ echo\n+ echo '## Containers @ testbed-node-1'\n+ echo\n+ osism container testbed-node-1 ps\nCONTAINER ID   IMAGE                                                                COMMAND                  CREATED          STATUS                    PORTS     NAMES\nb96d6747a23c   registry.osism.tech/kolla/octavia-worker:2024.2                      \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_worker\ne13afae5aa0a   registry.osism.tech/kolla/octavia-housekeeping:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_housekeeping\nda03c3f203d8   registry.osism.tech/kolla/octavia-health-manager:2024.2              \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_health_manager\n9cfb28dcd39f   registry.osism.tech/kolla/octavia-driver-agent:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes                        octavia_driver_agent\n62a640b1abe4   registry.osism.tech/kolla/octavia-api:2024.2                         \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_api\n42fbd84ca280   registry.osism.tech/kolla/nova-novncproxy:2024.2                     \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_novncproxy\nb5219b0892fb   registry.osism.tech/kolla/nova-conductor:2024.2                      \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_conductor\nd0994dd06228   registry.osism.tech/kolla/grafana:2024.2                             \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes                        grafana\nfa5a07dbd841   registry.osism.tech/kolla/nova-api:2024.2                            \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_api\n708a6fe44ce6   registry.osism.tech/kolla/nova-scheduler:2024.2                      \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_scheduler\nfc9c5048ab2c   registry.osism.tech/kolla/cinder-backup:2024.2                       \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_backup\n6c3ee47cae97   registry.osism.tech/kolla/cinder-volume:2024.2                       \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_volume\n078677f3587a   registry.osism.tech/kolla/glance-api:2024.2                          \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             glance_api\nb9fae0b403e3   registry.osism.tech/kolla/cinder-scheduler:2024.2                    \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_scheduler\n5683d8da194d   registry.osism.tech/kolla/cinder-api:2024.2                          \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_api\nc671e3dd9cbb   registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2   \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_elasticsearch_exporter\n43fd6bbce2aa   registry.osism.tech/kolla/prometheus-cadvisor:2024.2                 \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_cadvisor\nd9fc738086bb   registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_memcached_exporter\n9761154a7e32   registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2          \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_mysqld_exporter\n54125ce3e08c   registry.osism.tech/kolla/prometheus-node-exporter:2024.2            \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_node_exporter\nd4a0b0e8dac8   registry.osism.tech/kolla/magnum-conductor:2024.2                    \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             magnum_conductor\n3e1bb59dea40   registry.osism.tech/kolla/magnum-api:2024.2                          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             magnum_api\n8ad6908d91f0   registry.osism.tech/kolla/neutron-server:2024.2                      \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             neutron_server\n12d4f9f919c3   registry.osism.tech/kolla/designate-worker:2024.2                    \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             designate_worker\n21f63530e59c   registry.osism.tech/kolla/placement-api:2024.2                       \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             placement_api\n40eeee9a200e   registry.osism.tech/kolla/designate-mdns:2024.2                      \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_mdns\n086862740b99   registry.osism.tech/kolla/designate-producer:2024.2                  \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_producer\n4c7bdbda027f   registry.osism.tech/kolla/designate-central:2024.2                   \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_central\n79982be92a5e   registry.osism.tech/kolla/designate-api:2024.2                       \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_api\n58f711bf95fb   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mgr -\u2026\"   14 minutes ago   Up 14 minutes                       ceph-mgr-testbed-node-1\n0aa25d6997ac   registry.osism.tech/kolla/barbican-worker:2024.2                     \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             barbican_worker\ne0df55fbd5d2   registry.osism.tech/kolla/designate-backend-bind9:2024.2             \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_backend_bind9\ne4902252a4d7   registry.osism.tech/kolla/barbican-keystone-listener:2024.2          \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             barbican_keystone_listener\n2ee84b1a3de7   registry.osism.tech/kolla/barbican-api:2024.2                        \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             barbican_api\nfad8b7e4e4f8   registry.osism.tech/kolla/keystone:2024.2                            \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             keystone\nbe98159212c6   registry.osism.tech/kolla/keystone-fernet:2024.2                     \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             keystone_fernet\n8391e6c5b1b4   registry.osism.tech/kolla/horizon:2024.2                             \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             horizon\n026edb57c8f2   registry.osism.tech/kolla/keystone-ssh:2024.2                        \"dumb-init --single-\u2026\"   18 minutes ago   Up 17 minutes (healthy)             keystone_ssh\ne3d0530f6f13   registry.osism.tech/kolla/opensearch-dashboards:2024.2               \"dumb-init --single-\u2026\"   19 minutes ago   Up 19 minutes (healthy)             opensearch_dashboards\n903dca551e29   registry.osism.tech/kolla/mariadb-server:2024.2                      \"dumb-init -- kolla_\u2026\"   20 minutes ago   Up 20 minutes (healthy)             mariadb\nabef70983a85   registry.osism.tech/kolla/opensearch:2024.2                          \"dumb-init --single-\u2026\"   20 minutes ago   Up 20 minutes (healthy)             opensearch\n4caa03a44222   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-crash\"    21 minutes ago   Up 21 minutes                       ceph-crash-testbed-node-1\nbd27d8a16f20   registry.osism.tech/kolla/keepalived:2024.2                          \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes                       keepalived\n76f86a34f0e0   registry.osism.tech/kolla/proxysql:2024.2                            \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes (healthy)             proxysql\n6deaa4399b35   registry.osism.tech/kolla/haproxy:2024.2                             \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes (healthy)             haproxy\n2e8697946763   registry.osism.tech/kolla/ovn-northd:2024.2                          \"dumb-init --single-\u2026\"   25 minutes ago   Up 24 minutes                       ovn_northd\nafc85eabf323   registry.osism.tech/kolla/ovn-sb-db-server:2024.2                    \"dumb-init --single-\u2026\"   25 minutes ago   Up 24 minutes                       ovn_sb_db\n5c29f22f1712   registry.osism.tech/kolla/ovn-nb-db-server:2024.2                    \"dumb-init --single-\u2026\"   25 minutes ago   Up 24 minutes                       ovn_nb_db\n2c89399c74be   registry.osism.tech/kolla/ovn-controller:2024.2                      \"dumb-init --single-\u2026\"   25 minutes ago   Up 25 minutes                       ovn_controller\nee2851244d11   registry.osism.tech/kolla/rabbitmq:2024.2                            \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes (healthy)             rabbitmq\na96b2c5299e5   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mon -\u2026\"   26 minutes ago   Up 26 minutes                       ceph-mon-testbed-node-1\nf7f9ce0f9b1c   registry.osism.tech/kolla/openvswitch-vswitchd:2024.2                \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             openvswitch_vswitchd\n3f280ab09ce9   registry.osism.tech/kolla/openvswitch-db-server:2024.2               \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             openvswitch_db\n1077e37f68a8   registry.osism.tech/kolla/redis-sentinel:2024.2                      \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             redis_sentinel\n8fb396a44b47   registry.osism.tech/kolla/redis:2024.2                               \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             redis\n2ed0203510f4   registry.osism.tech/kolla/memcached:2024.2                           \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             memcached\n7631a1302929   registry.osism.tech/kolla/cron:2024.2                                \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes                       cron\n0399bb27a879   registry.osism.tech/kolla/kolla-toolbox:2024.2                       \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes                       kolla_toolbox\n6618cc36166a   registry.osism.tech/kolla/fluentd:2024.2                             \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                       fluentd\n\n## Images @ testbed-node-1\n\n+ echo\n+ echo '## Images @ testbed-node-1'\n+ echo\n+ osism container testbed-node-1 images\nREPOSITORY                                                    TAG       IMAGE ID       CREATED        SIZE\nregistry.osism.tech/osism/ceph-daemon                         reef      60d930974fb1   22 hours ago   1.27GB\nregistry.osism.tech/kolla/keepalived                          2024.2    b2b59a47e345   23 hours ago   286MB\nregistry.osism.tech/kolla/grafana                             2024.2    25f6c06d0021   23 hours ago   1.04GB\nregistry.osism.tech/kolla/kolla-toolbox                       2024.2    2ff405629e62   23 hours ago   677MB\nregistry.osism.tech/kolla/cron                                2024.2    23ec8544c793   23 hours ago   275MB\nregistry.osism.tech/kolla/opensearch-dashboards               2024.2    687dcde2bcb8   23 hours ago   1.54GB\nregistry.osism.tech/kolla/opensearch                          2024.2    b409ad33e488   23 hours ago   1.57GB\nregistry.osism.tech/kolla/memcached                           2024.2    ed2bff2a6973   23 hours ago   276MB\nregistry.osism.tech/kolla/proxysql                            2024.2    b4d6db8fd341   23 hours ago   426MB\nregistry.osism.tech/kolla/fluentd                             2024.2    851a383a0141   23 hours ago   589MB\nregistry.osism.tech/kolla/rabbitmq                            2024.2    a4ab33fd70fd   23 hours ago   332MB\nregistry.osism.tech/kolla/haproxy                             2024.2    3fa2a10107a8   24 hours ago   284MB\nregistry.osism.tech/kolla/redis-sentinel                      2024.2    5f4f04f4383c   24 hours ago   282MB\nregistry.osism.tech/kolla/redis                               2024.2    533c487568ee   24 hours ago   282MB\nregistry.osism.tech/kolla/mariadb-server                      2024.2    297f8bb339a1   24 hours ago   462MB\nregistry.osism.tech/kolla/prometheus-memcached-exporter       2024.2    dd336c43b779   24 hours ago   308MB\nregistry.osism.tech/kolla/prometheus-node-exporter            2024.2    e9fbd7018fbf   24 hours ago   315MB\nregistry.osism.tech/kolla/prometheus-mysqld-exporter          2024.2    548293489d76   24 hours ago   310MB\nregistry.osism.tech/kolla/prometheus-cadvisor                 2024.2    1520e18eb877   24 hours ago   367MB\nregistry.osism.tech/kolla/prometheus-elasticsearch-exporter   2024.2    5b796764074a   24 hours ago   301MB\nregistry.osism.tech/kolla/horizon                             2024.2    50eeb61edd36   24 hours ago   1.16GB\nregistry.osism.tech/kolla/openvswitch-vswitchd                2024.2    dbcb0c63d3a2   24 hours ago   288MB\nregistry.osism.tech/kolla/openvswitch-db-server               2024.2    a8f2511eff1f   24 hours ago   288MB\nregistry.osism.tech/kolla/neutron-server                      2024.2    7c6811afdb1b   24 hours ago   1.17GB\nregistry.osism.tech/kolla/octavia-api                         2024.2    eaf97bcdf081   24 hours ago   1.06GB\nregistry.osism.tech/kolla/octavia-driver-agent                2024.2    4c4abbcba868   24 hours ago   1.06GB\nregistry.osism.tech/kolla/octavia-worker                      2024.2    ae8999de916c   24 hours ago   1.04GB\nregistry.osism.tech/kolla/octavia-housekeeping                2024.2    e15b03944567   24 hours ago   1.04GB\nregistry.osism.tech/kolla/octavia-health-manager              2024.2    41dcecaeaf89   24 hours ago   1.04GB\nregistry.osism.tech/kolla/keystone                            2024.2    a6b20f41325d   24 hours ago   1.08GB\nregistry.osism.tech/kolla/keystone-fernet                     2024.2    15aaf5c48263   24 hours ago   1.05GB\nregistry.osism.tech/kolla/keystone-ssh                        2024.2    2634cde67cfd   24 hours ago   1.05GB\nregistry.osism.tech/kolla/cinder-backup                       2024.2    5f8e22fdae27   24 hours ago   1.42GB\nregistry.osism.tech/kolla/cinder-volume                       2024.2    29801e980d03   24 hours ago   1.73GB\nregistry.osism.tech/kolla/cinder-scheduler                    2024.2    7d98f6182a5d   24 hours ago   1.41GB\nregistry.osism.tech/kolla/cinder-api                          2024.2    c16461b7e25c   24 hours ago   1.41GB\nregistry.osism.tech/kolla/barbican-worker                     2024.2    cbe9a09e642e   24 hours ago   1GB\nregistry.osism.tech/kolla/barbican-keystone-listener          2024.2    8b79c739c5aa   24 hours ago   1GB\nregistry.osism.tech/kolla/barbican-api                        2024.2    ec75be03f45f   24 hours ago   1e+03MB\nregistry.osism.tech/kolla/nova-scheduler                      2024.2    d13b758badcb   24 hours ago   1.22GB\nregistry.osism.tech/kolla/nova-api                            2024.2    fa74d9e1d80c   24 hours ago   1.22GB\nregistry.osism.tech/kolla/nova-novncproxy                     2024.2    b0f517fc2110   24 hours ago   1.38GB\nregistry.osism.tech/kolla/nova-conductor                      2024.2    cc5099f460a9   24 hours ago   1.22GB\nregistry.osism.tech/kolla/glance-api                          2024.2    0c1d9d7160de   24 hours ago   1.11GB\nregistry.osism.tech/kolla/magnum-conductor                    2024.2    e4d602a3ad04   24 hours ago   1.25GB\nregistry.osism.tech/kolla/magnum-api                          2024.2    083abdbaa2fa   24 hours ago   1.14GB\nregistry.osism.tech/kolla/placement-api                       2024.2    dfc171655fae   24 hours ago   985MB\nregistry.osism.tech/kolla/designate-worker                    2024.2    e9363cf6baa5   24 hours ago   998MB\nregistry.osism.tech/kolla/designate-producer                  2024.2    1d3f00ae0cb0   24 hours ago   994MB\nregistry.osism.tech/kolla/designate-mdns                      2024.2    b74dea85ff31   24 hours ago   994MB\nregistry.osism.tech/kolla/designate-backend-bind9             2024.2    71020f79d0b8   24 hours ago   998MB\nregistry.osism.tech/kolla/designate-central                   2024.2    e270fdf55dd9   24 hours ago   993MB\nregistry.osism.tech/kolla/designate-api                       2024.2    2b71cfe02281   24 hours ago   994MB\nregistry.osism.tech/kolla/ovn-sb-db-server                    2024.2    c3d787fcf696   24 hours ago   850MB\nregistry.osism.tech/kolla/ovn-nb-db-server                    2024.2    91492a927f10   24 hours ago   850MB\nregistry.osism.tech/kolla/ovn-controller                      2024.2    df4830242254   24 hours ago   850MB\nregistry.osism.tech/kolla/ovn-northd                          2024.2    521cb38ac60e   24 hours ago   850MB\n+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2\n++ semver latest 5.0.0\n\n## Containers @ testbed-node-2\n\n+ [[ -1 -eq -1 ]]\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ echo\n+ echo '## Containers @ testbed-node-2'\n+ echo\n+ osism container testbed-node-2 ps\nCONTAINER ID   IMAGE                                                                COMMAND                  CREATED          STATUS                    PORTS     NAMES\n0ae5e2e46239   registry.osism.tech/kolla/octavia-worker:2024.2                      \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_worker\nac59fbdcb5c0   registry.osism.tech/kolla/octavia-housekeeping:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_housekeeping\n38f548bbc74b   registry.osism.tech/kolla/octavia-health-manager:2024.2              \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_health_manager\n0dcc5a926d5b   registry.osism.tech/kolla/octavia-driver-agent:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes                        octavia_driver_agent\n8a02191defe9   registry.osism.tech/kolla/octavia-api:2024.2                         \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_api\n39264437dc74   registry.osism.tech/kolla/nova-novncproxy:2024.2                     \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_novncproxy\nf42b2ff11f0b   registry.osism.tech/kolla/nova-conductor:2024.2                      \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_conductor\n18b5f3a218c9   registry.osism.tech/kolla/grafana:2024.2                             \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes                        grafana\n125c018b939c   registry.osism.tech/kolla/nova-api:2024.2                            \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_api\n40c1053d59c7   registry.osism.tech/kolla/nova-scheduler:2024.2                      \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_scheduler\nf7af2ca7b532   registry.osism.tech/kolla/cinder-backup:2024.2                       \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_backup\n69c2a2a2d036   registry.osism.tech/kolla/cinder-volume:2024.2                       \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_volume\na76993a8d64a   registry.osism.tech/kolla/glance-api:2024.2                          \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             glance_api\n56495cd3659a   registry.osism.tech/kolla/cinder-scheduler:2024.2                    \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_scheduler\n2f36c0d068ef   registry.osism.tech/kolla/cinder-api:2024.2                          \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_api\n704a31a41152   registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2   \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_elasticsearch_exporter\n42d0729cc93d   registry.osism.tech/kolla/prometheus-cadvisor:2024.2                 \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_cadvisor\n88896184888a   registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_memcached_exporter\n56ab80b6db9f   registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2          \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_mysqld_exporter\n32fb2e56f1e2   registry.osism.tech/kolla/prometheus-node-exporter:2024.2            \"dumb-init --single-\u2026\"   12 minutes ago   Up 11 minutes                       prometheus_node_exporter\nf1592e2057c5   registry.osism.tech/kolla/magnum-conductor:2024.2                    \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             magnum_conductor\n722b223d64fb   registry.osism.tech/kolla/magnum-api:2024.2                          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             magnum_api\ndb28a4c0beec   registry.osism.tech/kolla/neutron-server:2024.2                      \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             neutron_server\n967298c31eb9   registry.osism.tech/kolla/designate-worker:2024.2                    \"dumb-init --single-\u2026\"   14 minutes ago   Up 13 minutes (healthy)             designate_worker\n1b4e69352242   registry.osism.tech/kolla/placement-api:2024.2                       \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             placement_api\n61db02eb584f   registry.osism.tech/kolla/designate-mdns:2024.2                      \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_mdns\n65134717cd42   registry.osism.tech/kolla/designate-producer:2024.2                  \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_producer\neb4673f5b25f   registry.osism.tech/kolla/designate-central:2024.2                   \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_central\n0db26038f628   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mgr -\u2026\"   14 minutes ago   Up 14 minutes                       ceph-mgr-testbed-node-2\nd1ae168b9a8d   registry.osism.tech/kolla/designate-api:2024.2                       \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_api\n4a2814b312f1   registry.osism.tech/kolla/barbican-worker:2024.2                     \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             barbican_worker\ndfd755dc6faa   registry.osism.tech/kolla/designate-backend-bind9:2024.2             \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_backend_bind9\n0cb1920e7dd4   registry.osism.tech/kolla/barbican-keystone-listener:2024.2          \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             barbican_keystone_listener\n3882321b1615   registry.osism.tech/kolla/barbican-api:2024.2                        \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_api\n4ab3833b2116   registry.osism.tech/kolla/keystone:2024.2                            \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             keystone\nf1f648fe902c   registry.osism.tech/kolla/keystone-fernet:2024.2                     \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             keystone_fernet\nef461c30c337   registry.osism.tech/kolla/horizon:2024.2                             \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             horizon\nc234a31af7b6   registry.osism.tech/kolla/keystone-ssh:2024.2                        \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_ssh\n2169ee164e49   registry.osism.tech/kolla/opensearch-dashboards:2024.2               \"dumb-init --single-\u2026\"   19 minutes ago   Up 19 minutes (healthy)             opensearch_dashboards\n5a7def510321   registry.osism.tech/kolla/mariadb-server:2024.2                      \"dumb-init -- kolla_\u2026\"   20 minutes ago   Up 20 minutes (healthy)             mariadb\n35628559acd2   registry.osism.tech/kolla/opensearch:2024.2                          \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             opensearch\na0ddbb326162   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-crash\"    21 minutes ago   Up 21 minutes                       ceph-crash-testbed-node-2\ncb39873970ea   registry.osism.tech/kolla/keepalived:2024.2                          \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes                       keepalived\n76a08b821526   registry.osism.tech/kolla/proxysql:2024.2                            \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes (healthy)             proxysql\ne831ec38884e   registry.osism.tech/kolla/haproxy:2024.2                             \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes (healthy)             haproxy\n4ba5cc59fc04   registry.osism.tech/kolla/ovn-northd:2024.2                          \"dumb-init --single-\u2026\"   25 minutes ago   Up 24 minutes                       ovn_northd\n8e344c80c65a   registry.osism.tech/kolla/ovn-sb-db-server:2024.2                    \"dumb-init --single-\u2026\"   25 minutes ago   Up 24 minutes                       ovn_sb_db\n18ffee7ce1bd   registry.osism.tech/kolla/ovn-nb-db-server:2024.2                    \"dumb-init --single-\u2026\"   25 minutes ago   Up 25 minutes                       ovn_nb_db\n4ce96a10d15b   registry.osism.tech/kolla/rabbitmq:2024.2                            \"dumb-init --single-\u2026\"   25 minutes ago   Up 25 minutes (healthy)             rabbitmq\n2194c60e76b2   registry.osism.tech/kolla/ovn-controller:2024.2                      \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       ovn_controller\n57e0bd4375ae   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mon -\u2026\"   26 minutes ago   Up 26 minutes                       ceph-mon-testbed-node-2\n99acbc14ede4   registry.osism.tech/kolla/openvswitch-vswitchd:2024.2                \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             openvswitch_vswitchd\n9e5f2b62a875   registry.osism.tech/kolla/openvswitch-db-server:2024.2               \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             openvswitch_db\n648e20862b4d   registry.osism.tech/kolla/redis-sentinel:2024.2                      \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             redis_sentinel\na32a1543446c   registry.osism.tech/kolla/redis:2024.2                               \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             redis\n0b9bd0c5649d   registry.osism.tech/kolla/memcached:2024.2                           \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             memcached\n5a710b5a3f25   registry.osism.tech/kolla/cron:2024.2                                \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes                       cron\nb098600212de   registry.osism.tech/kolla/kolla-toolbox:2024.2                       \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes                       kolla_toolbox\nc921c720e066   registry.osism.tech/kolla/fluentd:2024.2                             \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                       fluentd\n\n## Images @ testbed-node-2\n\n+ echo\n+ echo '## Images @ testbed-node-2'\n+ echo\n+ osism container testbed-node-2 images\nREPOSITORY                                                    TAG       IMAGE ID       CREATED        SIZE\nregistry.osism.tech/osism/ceph-daemon                         reef      60d930974fb1   22 hours ago   1.27GB\nregistry.osism.tech/kolla/keepalived                          2024.2    b2b59a47e345   23 hours ago   286MB\nregistry.osism.tech/kolla/grafana                             2024.2    25f6c06d0021   23 hours ago   1.04GB\nregistry.osism.tech/kolla/kolla-toolbox                       2024.2    2ff405629e62   23 hours ago   677MB\nregistry.osism.tech/kolla/cron                                2024.2    23ec8544c793   23 hours ago   275MB\nregistry.osism.tech/kolla/opensearch-dashboards               2024.2    687dcde2bcb8   23 hours ago   1.54GB\nregistry.osism.tech/kolla/opensearch                          2024.2    b409ad33e488   23 hours ago   1.57GB\nregistry.osism.tech/kolla/memcached                           2024.2    ed2bff2a6973   23 hours ago   276MB\nregistry.osism.tech/kolla/proxysql                            2024.2    b4d6db8fd341   24 hours ago   426MB\nregistry.osism.tech/kolla/fluentd                             2024.2    851a383a0141   24 hours ago   589MB\nregistry.osism.tech/kolla/rabbitmq                            2024.2    a4ab33fd70fd   24 hours ago   332MB\nregistry.osism.tech/kolla/haproxy                             2024.2    3fa2a10107a8   24 hours ago   284MB\nregistry.osism.tech/kolla/redis-sentinel                      2024.2    5f4f04f4383c   24 hours ago   282MB\nregistry.osism.tech/kolla/redis                               2024.2    533c487568ee   24 hours ago   282MB\nregistry.osism.tech/kolla/mariadb-server                      2024.2    297f8bb339a1   24 hours ago   462MB\nregistry.osism.tech/kolla/prometheus-memcached-exporter       2024.2    dd336c43b779   24 hours ago   308MB\nregistry.osism.tech/kolla/prometheus-node-exporter            2024.2    e9fbd7018fbf   24 hours ago   315MB\nregistry.osism.tech/kolla/prometheus-mysqld-exporter          2024.2    548293489d76   24 hours ago   310MB\nregistry.osism.tech/kolla/prometheus-cadvisor                 2024.2    1520e18eb877   24 hours ago   367MB\nregistry.osism.tech/kolla/prometheus-elasticsearch-exporter   2024.2    5b796764074a   24 hours ago   301MB\nregistry.osism.tech/kolla/horizon                             2024.2    50eeb61edd36   24 hours ago   1.16GB\nregistry.osism.tech/kolla/openvswitch-vswitchd                2024.2    dbcb0c63d3a2   24 hours ago   288MB\nregistry.osism.tech/kolla/openvswitch-db-server               2024.2    a8f2511eff1f   24 hours ago   288MB\nregistry.osism.tech/kolla/neutron-server                      2024.2    7c6811afdb1b   24 hours ago   1.17GB\nregistry.osism.tech/kolla/octavia-api                         2024.2    eaf97bcdf081   24 hours ago   1.06GB\nregistry.osism.tech/kolla/octavia-driver-agent                2024.2    4c4abbcba868   24 hours ago   1.06GB\nregistry.osism.tech/kolla/octavia-worker                      2024.2    ae8999de916c   24 hours ago   1.04GB\nregistry.osism.tech/kolla/octavia-housekeeping                2024.2    e15b03944567   24 hours ago   1.04GB\nregistry.osism.tech/kolla/octavia-health-manager              2024.2    41dcecaeaf89   24 hours ago   1.04GB\nregistry.osism.tech/kolla/keystone                            2024.2    a6b20f41325d   24 hours ago   1.08GB\nregistry.osism.tech/kolla/keystone-fernet                     2024.2    15aaf5c48263   24 hours ago   1.05GB\nregistry.osism.tech/kolla/keystone-ssh                        2024.2    2634cde67cfd   24 hours ago   1.05GB\nregistry.osism.tech/kolla/cinder-backup                       2024.2    5f8e22fdae27   24 hours ago   1.42GB\nregistry.osism.tech/kolla/cinder-volume                       2024.2    29801e980d03   24 hours ago   1.73GB\nregistry.osism.tech/kolla/cinder-scheduler                    2024.2    7d98f6182a5d   24 hours ago   1.41GB\nregistry.osism.tech/kolla/cinder-api                          2024.2    c16461b7e25c   24 hours ago   1.41GB\nregistry.osism.tech/kolla/barbican-worker                     2024.2    cbe9a09e642e   24 hours ago   1GB\nregistry.osism.tech/kolla/barbican-keystone-listener          2024.2    8b79c739c5aa   24 hours ago   1GB\nregistry.osism.tech/kolla/barbican-api                        2024.2    ec75be03f45f   24 hours ago   1e+03MB\nregistry.osism.tech/kolla/nova-scheduler                      2024.2    d13b758badcb   24 hours ago   1.22GB\nregistry.osism.tech/kolla/nova-api                            2024.2    fa74d9e1d80c   24 hours ago   1.22GB\nregistry.osism.tech/kolla/nova-novncproxy                     2024.2    b0f517fc2110   24 hours ago   1.38GB\nregistry.osism.tech/kolla/nova-conductor                      2024.2    cc5099f460a9   24 hours ago   1.22GB\nregistry.osism.tech/kolla/glance-api                          2024.2    0c1d9d7160de   24 hours ago   1.11GB\nregistry.osism.tech/kolla/magnum-conductor                    2024.2    e4d602a3ad04   24 hours ago   1.25GB\nregistry.osism.tech/kolla/magnum-api                          2024.2    083abdbaa2fa   24 hours ago   1.14GB\nregistry.osism.tech/kolla/placement-api                       2024.2    dfc171655fae   24 hours ago   985MB\nregistry.osism.tech/kolla/designate-worker                    2024.2    e9363cf6baa5   24 hours ago   998MB\nregistry.osism.tech/kolla/designate-producer                  2024.2    1d3f00ae0cb0   24 hours ago   994MB\nregistry.osism.tech/kolla/designate-mdns                      2024.2    b74dea85ff31   24 hours ago   994MB\nregistry.osism.tech/kolla/designate-backend-bind9             2024.2    71020f79d0b8   24 hours ago   998MB\nregistry.osism.tech/kolla/designate-central                   2024.2    e270fdf55dd9   24 hours ago   993MB\nregistry.osism.tech/kolla/designate-api                       2024.2    2b71cfe02281   24 hours ago   994MB\nregistry.osism.tech/kolla/ovn-nb-db-server                    2024.2    91492a927f10   24 hours ago   850MB\nregistry.osism.tech/kolla/ovn-sb-db-server                    2024.2    c3d787fcf696   24 hours ago   850MB\nregistry.osism.tech/kolla/ovn-controller                      2024.2    df4830242254   24 hours ago   850MB\nregistry.osism.tech/kolla/ovn-northd                          2024.2    521cb38ac60e   24 hours ago   850MB\n+ sh -c /opt/configuration/scripts/check-services.sh\n+ set -e\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=true\n++ TEMPEST=true\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]\n+ sh -c /opt/configuration/scripts/check/100-ceph-with-ansible.sh\n+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/configuration/scripts/manager-version.sh\n+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml\n\n# Ceph status\n\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n+ echo\n+ echo '# Ceph status'\n+ echo\n+ ceph -s\n  cluster:\n    id:     11111111-1111-1111-1111-111111111111\n    health: HEALTH_OK\n \n  services:\n    mon: 3 daemons, quorum testbed-node-0,testbed-node-1,testbed-node-2 (age 26m)\n    mgr: testbed-node-1(active, since 14m), standbys: testbed-node-0, testbed-node-2\n    mds: 1/1 daemons up, 2 standby\n    osd: 6 osds: 6 up (since 22m), 6 in (since 23m)\n    rgw: 3 daemons active (3 hosts, 1 zones)\n \n  data:\n    volumes: 1/1 healthy\n    pools:   14 pools, 401 pgs\n    objects: 556 objects, 2.2 GiB\n    usage:   7.1 GiB used, 113 GiB / 120 GiB avail\n    pgs:     401 active+clean\n \n\n# Ceph versions\n\n+ echo\n+ echo '# Ceph versions'\n+ echo\n+ ceph versions\n{\n    \"mon\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3\n    },\n    \"mgr\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3\n    },\n    \"osd\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 6\n    },\n    \"mds\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3\n    },\n    \"rgw\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3\n    },\n    \"overall\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 18\n    }\n}\n\n# Ceph OSD tree\n\n+ echo\n+ echo '# Ceph OSD tree'\n+ echo\n+ ceph osd df tree\nID  CLASS  WEIGHT   REWEIGHT  SIZE     RAW USE   DATA     OMAP     META     AVAIL    %USE  VAR   PGS  STATUS  TYPE NAME              \n-1         0.11691         -  120 GiB   7.1 GiB  6.7 GiB    6 KiB  421 MiB  113 GiB  5.91  1.00    -          root default           \n-3         0.03897         -   40 GiB   2.4 GiB  2.2 GiB    2 KiB  139 MiB   38 GiB  5.91  1.00    -              host testbed-node-3\n 0    hdd  0.01949   1.00000   20 GiB   1.0 GiB  955 MiB    1 KiB   70 MiB   19 GiB  5.01  0.85  189      up          osd.0          \n 3    hdd  0.01949   1.00000   20 GiB   1.4 GiB  1.3 GiB    1 KiB   70 MiB   19 GiB  6.81  1.15  201      up          osd.3          \n-5         0.03897         -   40 GiB   2.4 GiB  2.2 GiB    2 KiB  143 MiB   38 GiB  5.92  1.00    -              host testbed-node-4\n 1    hdd  0.01949   1.00000   20 GiB  1009 MiB  940 MiB    1 KiB   70 MiB   19 GiB  4.93  0.83  195      up          osd.1          \n 5    hdd  0.01949   1.00000   20 GiB   1.4 GiB  1.3 GiB    1 KiB   74 MiB   19 GiB  6.90  1.17  197      up          osd.5          \n-7         0.03897         -   40 GiB   2.4 GiB  2.2 GiB    2 KiB  139 MiB   38 GiB  5.91  1.00    -              host testbed-node-5\n 2    hdd  0.01949   1.00000   20 GiB   1.6 GiB  1.5 GiB    1 KiB   70 MiB   18 GiB  7.79  1.32  198      up          osd.2          \n 4    hdd  0.01949   1.00000   20 GiB   825 MiB  755 MiB    1 KiB   70 MiB   19 GiB  4.03  0.68  190      up          osd.4          \n                       TOTAL  120 GiB   7.1 GiB  6.7 GiB  9.3 KiB  421 MiB  113 GiB  5.91                                            \nMIN/MAX VAR: 0.68/1.32  STDDEV: 1.33\n\n# Ceph monitor status\n\n+ echo\n+ echo '# Ceph monitor status'\n+ echo\n+ ceph mon stat\ne1: 3 mons at {testbed-node-0=[v2:192.168.16.10:3300/0,v1:192.168.16.10:6789/0],testbed-node-1=[v2:192.168.16.11:3300/0,v1:192.168.16.11:6789/0],testbed-node-2=[v2:192.168.16.12:3300/0,v1:192.168.16.12:6789/0]} removed_ranks: {} disallowed_leaders: {}, election epoch 8, leader 0 testbed-node-0, quorum 0,1,2 testbed-node-0,testbed-node-1,testbed-node-2\n\n# Ceph quorum status\n\n+ echo\n+ echo '# Ceph quorum status'\n+ echo\n+ jq\n+ ceph quorum_status\n{\n  \"election_epoch\": 8,\n  \"quorum\": [\n    0,\n    1,\n    2\n  ],\n  \"quorum_names\": [\n    \"testbed-node-0\",\n    \"testbed-node-1\",\n    \"testbed-node-2\"\n  ],\n  \"quorum_leader_name\": \"testbed-node-0\",\n  \"quorum_age\": 1582,\n  \"features\": {\n    \"quorum_con\": \"4540138322906710015\",\n    \"quorum_mon\": [\n      \"kraken\",\n      \"luminous\",\n      \"mimic\",\n      \"osdmap-prune\",\n      \"nautilus\",\n      \"octopus\",\n      \"pacific\",\n      \"elector-pinging\",\n      \"quincy\",\n      \"reef\"\n    ]\n  },\n  \"monmap\": {\n    \"epoch\": 1,\n    \"fsid\": \"11111111-1111-1111-1111-111111111111\",\n    \"modified\": \"2026-03-21T00:48:46.754580Z\",\n    \"created\": \"2026-03-21T00:48:46.754580Z\",\n    \"min_mon_release\": 18,\n    \"min_mon_release_name\": \"reef\",\n    \"election_strategy\": 1,\n    \"disallowed_leaders: \": \"\",\n    \"stretch_mode\": false,\n    \"tiebreaker_mon\": \"\",\n    \"removed_ranks: \": \"\",\n    \"features\": {\n      \"persistent\": [\n        \"kraken\",\n        \"luminous\",\n        \"mimic\",\n        \"osdmap-prune\",\n        \"nautilus\",\n        \"octopus\",\n        \"pacific\",\n        \"elector-pinging\",\n        \"quincy\",\n        \"reef\"\n      ],\n      \"optional\": []\n    },\n    \"mons\": [\n      {\n        \"rank\": 0,\n        \"name\": \"testbed-node-0\",\n        \"public_addrs\": {\n          \"addrvec\": [\n            {\n              \"type\": \"v2\",\n              \"addr\": \"192.168.16.10:3300\",\n              \"nonce\": 0\n            },\n            {\n              \"type\": \"v1\",\n              \"addr\": \"192.168.16.10:6789\",\n              \"nonce\": 0\n            }\n          ]\n        },\n        \"addr\": \"192.168.16.10:6789/0\",\n        \"public_addr\": \"192.168.16.10:6789/0\",\n        \"priority\": 0,\n        \"weight\": 0,\n        \"crush_location\": \"{}\"\n      },\n      {\n        \"rank\": 1,\n        \"name\": \"testbed-node-1\",\n        \"public_addrs\": {\n          \"addrvec\": [\n            {\n              \"type\": \"v2\",\n              \"addr\": \"192.168.16.11:3300\",\n              \"nonce\": 0\n            },\n            {\n              \"type\": \"v1\",\n              \"addr\": \"192.168.16.11:6789\",\n              \"nonce\": 0\n            }\n          ]\n        },\n        \"addr\": \"192.168.16.11:6789/0\",\n        \"public_addr\": \"192.168.16.11:6789/0\",\n        \"priority\": 0,\n        \"weight\": 0,\n        \"crush_location\": \"{}\"\n      },\n      {\n        \"rank\": 2,\n        \"name\": \"testbed-node-2\",\n        \"public_addrs\": {\n          \"addrvec\": [\n            {\n              \"type\": \"v2\",\n              \"addr\": \"192.168.16.12:3300\",\n              \"nonce\": 0\n            },\n            {\n              \"type\": \"v1\",\n              \"addr\": \"192.168.16.12:6789\",\n              \"nonce\": 0\n            }\n          ]\n        },\n        \"addr\": \"192.168.16.12:6789/0\",\n        \"public_addr\": \"192.168.16.12:6789/0\",\n        \"priority\": 0,\n        \"weight\": 0,\n        \"crush_location\": \"{}\"\n      }\n    ]\n  }\n}\n\n# Ceph free space status\n\n+ echo\n+ echo '# Ceph free space status'\n+ echo\n+ ceph df\n--- RAW STORAGE ---\nCLASS     SIZE    AVAIL     USED  RAW USED  %RAW USED\nhdd    120 GiB  113 GiB  7.1 GiB   7.1 GiB       5.91\nTOTAL  120 GiB  113 GiB  7.1 GiB   7.1 GiB       5.91\n \n--- POOLS ---\nPOOL                       ID  PGS   STORED  OBJECTS     USED  %USED  MAX AVAIL\n.mgr                        1    1  577 KiB        2  1.1 MiB      0     52 GiB\ncephfs_data                 2   32      0 B        0      0 B      0     35 GiB\ncephfs_metadata             3   16  4.4 KiB       22   96 KiB      0     35 GiB\ndefault.rgw.buckets.data    4   32      0 B        0      0 B      0     35 GiB\ndefault.rgw.buckets.index   5   32      0 B        0      0 B      0     35 GiB\ndefault.rgw.control         6   32      0 B        8      0 B      0     35 GiB\ndefault.rgw.log             7   32  3.6 KiB      209  408 KiB      0     35 GiB\ndefault.rgw.meta            8   32      0 B        0      0 B      0     35 GiB\n.rgw.root                   9   32  3.9 KiB        8   64 KiB      0     52 GiB\nbackups                    10   32     19 B        2   12 KiB      0     35 GiB\nvolumes                    11   32     19 B        2   12 KiB      0     35 GiB\nimages                     12   32  2.2 GiB      299  6.7 GiB   5.99     35 GiB\nmetrics                    13   32     19 B        2   12 KiB      0     35 GiB\nvms                        14   32     19 B        2   12 KiB      0     35 GiB\n++ semver latest 5.0.0\n+ [[ -1 -eq -1 ]]\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ [[ ! -e /etc/redhat-release ]]\n+ osism apply facts\n\u001b[32m2026-03-21 01:15:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of facts.\u001b[0m\n\u001b[32m2026-03-21 01:15:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 05981b19-0a4e-4355-8129-9611087a39d0 (facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 01:15:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 05981b19-0a4e-4355-8129-9611087a39d0 (facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role facts] ********************************************************\n\nTASK [osism.commons.facts : Create custom facts directory] *********************\nSaturday 21 March 2026  01:15:48 +0000 (0:00:00.347)       0:00:00.347 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.facts : Copy fact files] ***********************************\nSaturday 21 March 2026  01:15:49 +0000 (0:00:01.413)       0:00:01.761 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nSaturday 21 March 2026  01:15:50 +0000 (0:00:01.325)       0:00:03.086 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nPLAY [Gather facts for all hosts if using --limit] *****************************\n\nTASK [Gather facts for all hosts] **********************************************\nSaturday 21 March 2026  01:15:55 +0000 (0:00:04.212)       0:00:07.298 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  01:15:55 +0000 (0:00:00.705)       0:00:08.004 ******** \n=============================================================================== \nGathers facts about hosts ----------------------------------------------- 4.21s\nosism.commons.facts : Create custom facts directory --------------------- 1.41s\nosism.commons.facts : Copy fact files ----------------------------------- 1.33s\nGather facts for all hosts ---------------------------------------------- 0.71s\n+ osism validate ceph-mons\n\nPLAY [Ceph validate mons] ******************************************************\n\nTASK [Get timestamp for report file] *******************************************\nSaturday 21 March 2026  01:16:11 +0000 (0:00:00.540)       0:00:00.541 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Create report output directory] ******************************************\nSaturday 21 March 2026  01:16:12 +0000 (0:00:00.988)       0:00:01.529 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Define report vars] ******************************************************\nSaturday 21 March 2026  01:16:12 +0000 (0:00:00.692)       0:00:02.221 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Prepare test data for container existance test] **************************\nSaturday 21 March 2026  01:16:12 +0000 (0:00:00.135)       0:00:02.357 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Get container info] ******************************************************\nSaturday 21 March 2026  01:16:13 +0000 (0:00:00.284)       0:00:02.642 ******** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Set test result to failed if container is missing] ***********************\nSaturday 21 March 2026  01:16:14 +0000 (0:00:01.359)       0:00:04.002 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to passed if container is existing] **********************\nSaturday 21 March 2026  01:16:14 +0000 (0:00:00.279)       0:00:04.282 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nSaturday 21 March 2026  01:16:15 +0000 (0:00:00.326)       0:00:04.608 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to failed if ceph-mon is not running] ********************\nSaturday 21 March 2026  01:16:15 +0000 (0:00:00.309)       0:00:04.918 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to passed if ceph-mon is running] ************************\nSaturday 21 March 2026  01:16:15 +0000 (0:00:00.435)       0:00:05.353 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nSaturday 21 March 2026  01:16:16 +0000 (0:00:00.297)       0:00:05.651 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nSaturday 21 March 2026  01:16:16 +0000 (0:00:00.239)       0:00:05.891 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nSaturday 21 March 2026  01:16:16 +0000 (0:00:00.241)       0:00:06.132 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nSaturday 21 March 2026  01:16:16 +0000 (0:00:00.238)       0:00:06.371 ******** \n\nTASK [Flush handlers] **********************************************************\nSaturday 21 March 2026  01:16:17 +0000 (0:00:00.081)       0:00:06.452 ******** \n\nTASK [Flush handlers] **********************************************************\nSaturday 21 March 2026  01:16:17 +0000 (0:00:00.069)       0:00:06.522 ******** \n\nTASK [Print report file information] *******************************************\nSaturday 21 March 2026  01:16:17 +0000 (0:00:00.216)       0:00:06.738 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Fail due to missing containers] ******************************************\nSaturday 21 March 2026  01:16:17 +0000 (0:00:00.254)       0:00:06.993 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Prepare quorum test vars] ************************************************\nSaturday 21 March 2026  01:16:17 +0000 (0:00:00.234)       0:00:07.228 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Get monmap info from one mon container] **********************************\nSaturday 21 March 2026  01:16:17 +0000 (0:00:00.128)       0:00:07.357 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [Set quorum test data] ****************************************************\nSaturday 21 March 2026  01:16:19 +0000 (0:00:01.946)       0:00:09.303 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail quorum test if not all monitors are in quorum] **********************\nSaturday 21 March 2026  01:16:20 +0000 (0:00:00.301)       0:00:09.605 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass quorum test if all monitors are in quorum] **************************\nSaturday 21 March 2026  01:16:20 +0000 (0:00:00.131)       0:00:09.737 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Set fsid test vars] ******************************************************\nSaturday 21 March 2026  01:16:20 +0000 (0:00:00.330)       0:00:10.067 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail Cluster FSID test if FSID does not match configuration] *************\nSaturday 21 March 2026  01:16:20 +0000 (0:00:00.285)       0:00:10.353 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass Cluster FSID test if it matches configuration] **********************\nSaturday 21 March 2026  01:16:21 +0000 (0:00:00.125)       0:00:10.479 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Prepare status test vars] ************************************************\nSaturday 21 March 2026  01:16:21 +0000 (0:00:00.124)       0:00:10.604 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Gather status data] ******************************************************\nSaturday 21 March 2026  01:16:21 +0000 (0:00:00.288)       0:00:10.892 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [Set health test data] ****************************************************\nSaturday 21 March 2026  01:16:23 +0000 (0:00:01.556)       0:00:12.449 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail cluster-health if health is not acceptable] *************************\nSaturday 21 March 2026  01:16:23 +0000 (0:00:00.318)       0:00:12.768 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass cluster-health if health is acceptable] *****************************\nSaturday 21 March 2026  01:16:23 +0000 (0:00:00.132)       0:00:12.901 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail cluster-health if health is not acceptable (strict)] ****************\nSaturday 21 March 2026  01:16:23 +0000 (0:00:00.142)       0:00:13.044 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass cluster-health if status is OK (strict)] ****************************\nSaturday 21 March 2026  01:16:23 +0000 (0:00:00.144)       0:00:13.188 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Set validation result to passed if no test failed] ***********************\nSaturday 21 March 2026  01:16:23 +0000 (0:00:00.135)       0:00:13.323 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Set validation result to failed if a test failed] ************************\nSaturday 21 March 2026  01:16:24 +0000 (0:00:00.260)       0:00:13.584 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nSaturday 21 March 2026  01:16:24 +0000 (0:00:00.247)       0:00:13.832 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nSaturday 21 March 2026  01:16:26 +0000 (0:00:01.746)       0:00:15.578 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nSaturday 21 March 2026  01:16:26 +0000 (0:00:00.250)       0:00:15.829 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nSaturday 21 March 2026  01:16:27 +0000 (0:00:00.645)       0:00:16.474 ******** \n\nTASK [Flush handlers] **********************************************************\nSaturday 21 March 2026  01:16:27 +0000 (0:00:00.070)       0:00:16.545 ******** \n\nTASK [Flush handlers] **********************************************************\nSaturday 21 March 2026  01:16:27 +0000 (0:00:00.076)       0:00:16.621 ******** \n\nRUNNING HANDLER [Write report file] ********************************************\nSaturday 21 March 2026  01:16:27 +0000 (0:00:00.076)       0:00:16.698 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Print report file information] *******************************************\nSaturday 21 March 2026  01:16:28 +0000 (0:00:01.248)       0:00:17.947 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => {\u001b[0m\n\u001b[0;32m    \"msg\": [\u001b[0m\n\u001b[0;32m        \"Validator run completed.\",\u001b[0m\n\u001b[0;32m        \"You can find the report file here:\",\u001b[0m\n\u001b[0;32m        \"/opt/reports/validator/ceph-mons-validator-2026-03-21T01:16:12+00:00-report.json\",\u001b[0m\n\u001b[0;32m        \"on the following host:\",\u001b[0m\n\u001b[0;32m        \"testbed-manager\"\u001b[0m\n\u001b[0;32m    ]\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=13  \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  01:16:28 +0000 (0:00:00.407)       0:00:18.354 ******** \n=============================================================================== \nGet monmap info from one mon container ---------------------------------- 1.95s\nAggregate test results step one ----------------------------------------- 1.75s\nGather status data ------------------------------------------------------ 1.56s\nGet container info ------------------------------------------------------ 1.36s\nWrite report file ------------------------------------------------------- 1.25s\nGet timestamp for report file ------------------------------------------- 0.99s\nCreate report output directory ------------------------------------------ 0.69s\nAggregate test results step three --------------------------------------- 0.65s\nSet test result to failed if ceph-mon is not running -------------------- 0.44s\nPrint report file information ------------------------------------------- 0.41s\nFlush handlers ---------------------------------------------------------- 0.37s\nPass quorum test if all monitors are in quorum -------------------------- 0.33s\nSet test result to passed if container is existing ---------------------- 0.33s\nSet health test data ---------------------------------------------------- 0.32s\nPrepare test data ------------------------------------------------------- 0.31s\nSet quorum test data ---------------------------------------------------- 0.30s\nSet test result to passed if ceph-mon is running ------------------------ 0.30s\nPrepare status test vars ------------------------------------------------ 0.29s\nSet fsid test vars ------------------------------------------------------ 0.29s\nPrepare test data for container existance test -------------------------- 0.28s\n+ osism validate ceph-mgrs\n\nPLAY [Ceph validate mgrs] ******************************************************\n\nTASK [Get timestamp for report file] *******************************************\nSaturday 21 March 2026  01:16:44 +0000 (0:00:00.506)       0:00:00.506 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Create report output directory] ******************************************\nSaturday 21 March 2026  01:16:45 +0000 (0:00:00.986)       0:00:01.492 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Define report vars] ******************************************************\nSaturday 21 March 2026  01:16:45 +0000 (0:00:00.688)       0:00:02.181 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Prepare test data for container existance test] **************************\nSaturday 21 March 2026  01:16:46 +0000 (0:00:00.126)       0:00:02.308 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Get container info] ******************************************************\nSaturday 21 March 2026  01:16:46 +0000 (0:00:00.276)       0:00:02.585 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to failed if container is missing] ***********************\nSaturday 21 March 2026  01:16:47 +0000 (0:00:01.382)       0:00:03.967 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to passed if container is existing] **********************\nSaturday 21 March 2026  01:16:48 +0000 (0:00:00.287)       0:00:04.254 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nSaturday 21 March 2026  01:16:48 +0000 (0:00:00.300)       0:00:04.554 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to failed if ceph-mgr is not running] ********************\nSaturday 21 March 2026  01:16:48 +0000 (0:00:00.300)       0:00:04.855 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to passed if ceph-mgr is running] ************************\nSaturday 21 March 2026  01:16:49 +0000 (0:00:00.457)       0:00:05.312 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nSaturday 21 March 2026  01:16:49 +0000 (0:00:00.290)       0:00:05.602 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nSaturday 21 March 2026  01:16:49 +0000 (0:00:00.233)       0:00:05.835 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nSaturday 21 March 2026  01:16:49 +0000 (0:00:00.262)       0:00:06.098 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nSaturday 21 March 2026  01:16:50 +0000 (0:00:00.237)       0:00:06.336 ******** \n\nTASK [Flush handlers] **********************************************************\nSaturday 21 March 2026  01:16:50 +0000 (0:00:00.094)       0:00:06.430 ******** \n\nTASK [Flush handlers] **********************************************************\nSaturday 21 March 2026  01:16:50 +0000 (0:00:00.076)       0:00:06.507 ******** \n\nTASK [Print report file information] *******************************************\nSaturday 21 March 2026  01:16:50 +0000 (0:00:00.235)       0:00:06.743 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Fail due to missing containers] ******************************************\nSaturday 21 March 2026  01:16:50 +0000 (0:00:00.262)       0:00:07.006 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Define mgr module test vars] *********************************************\nSaturday 21 March 2026  01:16:51 +0000 (0:00:00.271)       0:00:07.278 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Gather list of mgr modules] **********************************************\nSaturday 21 March 2026  01:16:51 +0000 (0:00:00.132)       0:00:07.410 ******** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [Parse mgr module list from json] *****************************************\nSaturday 21 March 2026  01:16:53 +0000 (0:00:02.099)       0:00:09.510 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Extract list of enabled mgr modules] *************************************\nSaturday 21 March 2026  01:16:53 +0000 (0:00:00.240)       0:00:09.751 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail test if mgr modules are disabled that should be enabled] ************\nSaturday 21 March 2026  01:16:53 +0000 (0:00:00.289)       0:00:10.041 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass test if required mgr modules are enabled] ***************************\nSaturday 21 March 2026  01:16:53 +0000 (0:00:00.114)       0:00:10.155 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Set validation result to passed if no test failed] ***********************\nSaturday 21 March 2026  01:16:54 +0000 (0:00:00.138)       0:00:10.294 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Set validation result to failed if a test failed] ************************\nSaturday 21 March 2026  01:16:54 +0000 (0:00:00.250)       0:00:10.544 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nSaturday 21 March 2026  01:16:54 +0000 (0:00:00.239)       0:00:10.784 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nSaturday 21 March 2026  01:16:56 +0000 (0:00:01.475)       0:00:12.260 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nSaturday 21 March 2026  01:16:56 +0000 (0:00:00.268)       0:00:12.528 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nSaturday 21 March 2026  01:16:56 +0000 (0:00:00.263)       0:00:12.792 ******** \n\nTASK [Flush handlers] **********************************************************\nSaturday 21 March 2026  01:16:56 +0000 (0:00:00.074)       0:00:12.866 ******** \n\nTASK [Flush handlers] **********************************************************\nSaturday 21 March 2026  01:16:56 +0000 (0:00:00.069)       0:00:12.936 ******** \n\nRUNNING HANDLER [Write report file] ********************************************\nSaturday 21 March 2026  01:16:56 +0000 (0:00:00.078)       0:00:13.014 ******** \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Print report file information] *******************************************\nSaturday 21 March 2026  01:16:58 +0000 (0:00:01.327)       0:00:14.342 ******** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => {\u001b[0m\n\u001b[0;32m    \"msg\": [\u001b[0m\n\u001b[0;32m        \"Validator run completed.\",\u001b[0m\n\u001b[0;32m        \"You can find the report file here:\",\u001b[0m\n\u001b[0;32m        \"/opt/reports/validator/ceph-mgrs-validator-2026-03-21T01:16:45+00:00-report.json\",\u001b[0m\n\u001b[0;32m        \"on the following host:\",\u001b[0m\n\u001b[0;32m        \"testbed-manager\"\u001b[0m\n\u001b[0;32m    ]\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  01:16:58 +0000 (0:00:00.392)       0:00:14.734 ******** \n=============================================================================== \nGather list of mgr modules ---------------------------------------------- 2.10s\nAggregate test results step one ----------------------------------------- 1.48s\nGet container info ------------------------------------------------------ 1.38s\nWrite report file ------------------------------------------------------- 1.33s\nGet timestamp for report file ------------------------------------------- 0.99s\nCreate report output directory ------------------------------------------ 0.69s\nSet test result to failed if ceph-mgr is not running -------------------- 0.46s\nFlush handlers ---------------------------------------------------------- 0.41s\nPrint report file information ------------------------------------------- 0.39s\nPrepare test data ------------------------------------------------------- 0.30s\nSet test result to passed if container is existing ---------------------- 0.30s\nSet test result to passed if ceph-mgr is running ------------------------ 0.29s\nExtract list of enabled mgr modules ------------------------------------- 0.29s\nSet test result to failed if container is missing ----------------------- 0.29s\nPrepare test data for container existance test -------------------------- 0.28s\nFail due to missing containers ------------------------------------------ 0.27s\nAggregate test results step two ----------------------------------------- 0.27s\nAggregate test results step three --------------------------------------- 0.26s\nPrint report file information ------------------------------------------- 0.26s\nAggregate test results step two ----------------------------------------- 0.26s\n+ osism validate ceph-osds\n\nPLAY [Ceph validate OSDs] ******************************************************\n\nTASK [Get timestamp for report file] *******************************************\nSaturday 21 March 2026  01:17:14 +0000 (0:00:00.497)       0:00:00.498 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nSaturday 21 March 2026  01:17:14 +0000 (0:00:00.949)       0:00:01.447 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Create report output directory] ******************************************\nSaturday 21 March 2026  01:17:15 +0000 (0:00:00.238)       0:00:01.686 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Define report vars] ******************************************************\nSaturday 21 March 2026  01:17:15 +0000 (0:00:00.672)       0:00:02.358 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Define OSD test variables] ***********************************************\nSaturday 21 March 2026  01:17:15 +0000 (0:00:00.123)       0:00:02.482 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate OSD devices for each host] *************************************\nSaturday 21 March 2026  01:17:16 +0000 (0:00:00.123)       0:00:02.605 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Define OSD test variables] ***********************************************\nSaturday 21 March 2026  01:17:16 +0000 (0:00:00.450)       0:00:03.056 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Calculate OSD devices for each host] *************************************\nSaturday 21 March 2026  01:17:16 +0000 (0:00:00.150)       0:00:03.206 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Calculate total number of OSDs in cluster] *******************************\nSaturday 21 March 2026  01:17:17 +0000 (0:00:00.356)       0:00:03.563 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nSaturday 21 March 2026  01:17:17 +0000 (0:00:00.325)       0:00:03.888 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get list of ceph-osd containers on host] *********************************\nSaturday 21 March 2026  01:17:17 +0000 (0:00:00.287)       0:00:04.176 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'b3834e4b9797d2ff4db9f4f433ff64037dedaa8e4e196486c83b6934b2b415da', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '04611561164cd51fc3a1ed5e9f60065177b9e751a138495a53804363d317566f', 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '686e42457e4a98d64855761766743009124da2d28c2111f4f56ec71ba08a5890', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '8a25dc8033b6564150316723eafb179e510067642a65fe7adc86d9e644ab9b14', 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 11 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '1ded55d684f98646e1b9581d31ebbf95ec0c9c8c87b69d8a1e3f6f77d9269b2a', 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 11 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '5a861065359ee5c8d9fcbe7cdcac9b4666265e7236c1deaa3d14cce10a7ebbb2', 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 12 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '0022450ab44c1664f8d92910e0b3377c16e5cf0f56e88032ea7b190c3d50d895', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 13 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '674e9590eb63fbaa5efe8cde9a89c3538774a40289a43b678d8aae81c688696e', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-rgw-default-testbed-node-3-rgw0', 'state': 'running', 'status': 'Up 20 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'fab7712f40b620b557b418bc8a67d66bfd9cab7ace760248fb3fc076f460282c', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-mds-testbed-node-3', 'state': 'running', 'status': 'Up 21 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'b6044d59fee87068d24f2a15531f68581ffb047be92817620778e740a75cc1c6', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-crash-testbed-node-3', 'state': 'running', 'status': 'Up 22 minutes'}) \u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'id': 'f80f9c86195d357ba685e4352428814027d0fab54202bb85f0680846813787de', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-3', 'state': 'running', 'status': 'Up 23 minutes'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'id': 'ced04366240e3aef9d90dc1cbb93c3ac0c4cee185f775b81dc73e7189c1d1e1c', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-0', 'state': 'running', 'status': 'Up 23 minutes'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '501b7122783ecf7a8d2a9445be6ef667b253ef950b196c910b800740ae2a1b5f', 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 26 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'ea950199405bc692edb293a5b3a5c718909ca184ab5d6a0dfd78b72ecfc51961', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 27 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '318388aef55ec0589d19a19ada3e0135c5c78e1a5f96c2573b650800443d3bc1', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 28 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '0ae70184f8566bb9deeafcbc6473423913b6bda2f6a90e5be8dbe3be275614ee', 'image': 'registry.osism.tech/kolla/cron:2024.2', 'name': '/cron', 'state': 'running', 'status': 'Up 28 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'bbf4cca6cfba90de64cadaa5eefae31d6081e3cf05b17890833266c229965fa5', 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 29 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '3820df04684d6d4bd30e1c4e426bfa66c74d53d8bbb13dd34b07f45716def56e', 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'name': '/fluentd', 'state': 'running', 'status': 'Up 29 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '4bcee744b0035069a448ebf926bbea6186f7d70ca7add1390f48951354d848ea', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '3f5e9d5cb4af813b127ff2510d0a21d593016f3a5e5f2b0255c6434e12546eb8', 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '82bb0e0b64eb13f425b2df3bcd5b458f4bdd35ea0c9092ec6ac83670dd2dbc4f', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '6000eeba069a7193758f7419466fd78c1dd986f8d0afab94b95b9f79577223d1', 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 11 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '1f6f3a4e0dd8c9d0780ae2a323f296b39901d8da51a2baf7fc96999ffe77469f', 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 11 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '33f610c36ac8458fce46755538cae737bc46001eaca9905e6d3f0e097662f9a4', 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 12 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'c691f7d217a9d8449cebbb687dbac850d02b0487365b2ad60434c3f6f589d8c6', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 12 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'f9709df7ba324efdcff3a8cf25d4bd1d1e79c443f1462e90429c0dfb0f219088', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-rgw-default-testbed-node-4-rgw0', 'state': 'running', 'status': 'Up 20 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '474c051b9924801f91e33a2c599e059a380ba100fa39ad81769345b1e9558599', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-mds-testbed-node-4', 'state': 'running', 'status': 'Up 21 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '8425dcc075d1c743c3594a66cc973c80cad5c730ef69eb67ce6327dee0a875a2', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-crash-testbed-node-4', 'state': 'running', 'status': 'Up 22 minutes'}) \u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'id': '9c825b9e0312f119368d2a0a7f03bd3dbf08b583add2062e8abd33e426500564', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-5', 'state': 'running', 'status': 'Up 23 minutes'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'id': '1add874758f001f932b28d061b71a9f39dc0f673ca85f86311e6920ca2db45b2', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-1', 'state': 'running', 'status': 'Up 23 minutes'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '670e4278f5d14145d6995bbc5b4d9aec7fa1d06776d8bc99c90a6e70214071e0', 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 26 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '822bf99a7617765d185192fc3c4880968cae2b43f98e55b2e4d4493d029f1826', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 27 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'd5f3c47c1de714009c246574488dea7b23dc8d2ef9bec0b35ac6032798555398', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 28 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '50aed0d9cdc3614469ef918fa6b903fff4cc2fa1103421c95ab18bb6c2e22c63', 'image': 'registry.osism.tech/kolla/cron:2024.2', 'name': '/cron', 'state': 'running', 'status': 'Up 28 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'ba8abe4b0317466dc1f636a0370fa0dcde2ef017dbd3caccb95ef5d13cc0f8fd', 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 29 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '9502548dd53e8359c6f99f0af769e6c2c54ae43bb11b57fc02653eeec7be811f', 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'name': '/fluentd', 'state': 'running', 'status': 'Up 29 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '4217b11e61a8f60f98ce804bc5bb30eb314cd4304461f7b63f1089f799acbb45', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '2b0e43585d8764f9c5c739ef4588b6b9ae88abcb577d0fed777c3508e40623db', 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '5e9ba54c5522ef2df374bab9292de1c7df20874ea099b1331c00ab0d9c062b80', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '1b8a7e9e094d41c11fd6bd46552a4366960e7ad6098a978d1310bb9a81fd1abb', 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 11 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '9b24a1153e98cbb631a896c63e9ee756923c22be26f175f91cc5dd8206f86797', 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 11 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '19601c0a2d6c2b10f8e644aebb2f1048678f0b6c0a8c037870184fc5ab9521a8', 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 12 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'b439d9b1776dcd4a9e555ce5a0065c128b6a9494afe441076937c4fac164db68', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 12 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '93a749a896acce9595ef8c691f29ca9095d7880856f41bfe32b6ca3286d598fa', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-rgw-default-testbed-node-5-rgw0', 'state': 'running', 'status': 'Up 20 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'fce8668785c102a8ae343c034b76ce0ee6f4155aaaff792ec316dd5fc935062f', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-mds-testbed-node-5', 'state': 'running', 'status': 'Up 21 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'c7183ffbca2bbe4fa711520a7c4f36a61fa21ca24441d0bceb49221e7d22077d', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-crash-testbed-node-5', 'state': 'running', 'status': 'Up 22 minutes'}) \u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'id': 'b39b31c03daf16fdb12e36656287c8ae5d75f5f32dec390e1a6b6fa6d8bf7a94', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-2', 'state': 'running', 'status': 'Up 23 minutes'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'id': '536e57de3f664f37f632d6d51267d8ead62a6fb4d1a06749969ddb6edf3119f4', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-4', 'state': 'running', 'status': 'Up 23 minutes'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '21301f5c0879a6affc5198ad889a1eb895d48ae7f32f3404a8e266c86d0be63b', 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 26 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '7c2076a2645cbf1f4568a36b52457a4b644600723a28afb1c29f09f5fece12c8', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 27 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '6696533b70d60f4d778334f6c795e1a3752c38b63ed2e20c33d28137394ebd73', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 28 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '15cc45b98c17825e6ac3d61d38e3cb9038433e22d41bba2f63922612822b9763', 'image': 'registry.osism.tech/kolla/cron:2024.2', 'name': '/cron', 'state': 'running', 'status': 'Up 28 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '4e853a7f8e10afbcc48c23e687d54fc607d86a9a6186d93057c82cac9c869f38', 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 29 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '5913715c59140ecf4be6c29f797c050fb5c792a82330d5983ac08105ef213efb', 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'name': '/fluentd', 'state': 'running', 'status': 'Up 29 minutes'}) \u001b[0m\n\nTASK [Get count of ceph-osd containers on host] ********************************\nSaturday 21 March 2026  01:17:18 +0000 (0:00:00.665)       0:00:04.841 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to failed when count of containers is wrong] *************\nSaturday 21 March 2026  01:17:18 +0000 (0:00:00.316)       0:00:05.158 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to passed if count matches] ******************************\nSaturday 21 March 2026  01:17:18 +0000 (0:00:00.294)       0:00:05.452 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nSaturday 21 March 2026  01:17:19 +0000 (0:00:00.333)       0:00:05.786 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get list of ceph-osd containers that are not running] ********************\nSaturday 21 March 2026  01:17:19 +0000 (0:00:00.424)       0:00:06.210 ******** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'ceph-osd-3', 'osd_id': '3', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'ceph-osd-0', 'osd_id': '0', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'ceph-osd-5', 'osd_id': '5', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'ceph-osd-1', 'osd_id': '1', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'ceph-osd-2', 'osd_id': '2', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'ceph-osd-4', 'osd_id': '4', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Get count of ceph-osd containers that are not running] *******************\nSaturday 21 March 2026  01:17:20 +0000 (0:00:00.322)       0:00:06.533 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to failed if an OSD is not running] **********************\nSaturday 21 March 2026  01:17:20 +0000 (0:00:00.284)       0:00:06.817 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to failed if an OSD is not running] **********************\nSaturday 21 March 2026  01:17:20 +0000 (0:00:00.293)       0:00:07.111 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to passed if all containers are running] *****************\nSaturday 21 March 2026  01:17:21 +0000 (0:00:00.480)       0:00:07.592 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nSaturday 21 March 2026  01:17:21 +0000 (0:00:00.284)       0:00:07.876 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nSaturday 21 March 2026  01:17:21 +0000 (0:00:00.241)       0:00:08.118 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nSaturday 21 March 2026  01:17:21 +0000 (0:00:00.249)       0:00:08.368 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nSaturday 21 March 2026  01:17:22 +0000 (0:00:00.239)       0:00:08.607 ******** \n\nTASK [Flush handlers] **********************************************************\nSaturday 21 March 2026  01:17:22 +0000 (0:00:00.068)       0:00:08.676 ******** \n\nTASK [Flush handlers] **********************************************************\nSaturday 21 March 2026  01:17:22 +0000 (0:00:00.067)       0:00:08.743 ******** \n\nTASK [Print report file information] *******************************************\nSaturday 21 March 2026  01:17:22 +0000 (0:00:00.068)       0:00:08.811 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail early due to containers not running] ********************************\nSaturday 21 March 2026  01:17:22 +0000 (0:00:00.593)       0:00:09.405 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nSaturday 21 March 2026  01:17:23 +0000 (0:00:00.247)       0:00:09.652 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set _mon_hostname fact] **************************************************\nSaturday 21 March 2026  01:17:23 +0000 (0:00:00.305)       0:00:09.957 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Get ceph osd tree] *******************************************************\nSaturday 21 March 2026  01:17:23 +0000 (0:00:00.228)       0:00:10.186 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [Parse osd tree from JSON] ************************************************\nSaturday 21 March 2026  01:17:25 +0000 (0:00:02.012)       0:00:12.198 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Get OSDs that are not up or in] ******************************************\nSaturday 21 March 2026  01:17:25 +0000 (0:00:00.116)       0:00:12.315 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Fail test if OSDs are not up or in] **************************************\nSaturday 21 March 2026  01:17:26 +0000 (0:00:00.280)       0:00:12.595 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Pass test if OSDs are all up and in] *************************************\nSaturday 21 March 2026  01:17:26 +0000 (0:00:00.113)       0:00:12.709 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nSaturday 21 March 2026  01:17:26 +0000 (0:00:00.126)       0:00:12.836 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [List ceph LVM volumes and collect data] **********************************\nSaturday 21 March 2026  01:17:26 +0000 (0:00:00.439)       0:00:13.276 ******** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Parse LVM data as JSON] **************************************************\nSaturday 21 March 2026  01:17:29 +0000 (0:00:02.505)       0:00:15.782 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get unencrypted and encrypted OSDs] **************************************\nSaturday 21 March 2026  01:17:29 +0000 (0:00:00.322)       0:00:16.105 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Fail if count of encrypted OSDs does not match] **************************\nSaturday 21 March 2026  01:17:30 +0000 (0:00:00.450)       0:00:16.555 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Pass if count of encrypted OSDs equals count of OSDs] ********************\nSaturday 21 March 2026  01:17:30 +0000 (0:00:00.473)       0:00:17.028 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Fail if count of unencrypted OSDs does not match] ************************\nSaturday 21 March 2026  01:17:30 +0000 (0:00:00.303)       0:00:17.331 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Pass if count of unencrypted OSDs equals count of OSDs] ******************\nSaturday 21 March 2026  01:17:31 +0000 (0:00:00.283)       0:00:17.615 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nSaturday 21 March 2026  01:17:31 +0000 (0:00:00.435)       0:00:18.051 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get CRUSH node data of each OSD host and root node childs] ***************\nSaturday 21 March 2026  01:17:32 +0000 (0:00:00.485)       0:00:18.536 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Calculate sub test expression results] ***********************************\nSaturday 21 March 2026  01:17:32 +0000 (0:00:00.460)       0:00:18.997 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Fail test if any sub test failed] ****************************************\nSaturday 21 March 2026  01:17:32 +0000 (0:00:00.294)       0:00:19.292 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Pass test if no sub test failed] *****************************************\nSaturday 21 March 2026  01:17:33 +0000 (0:00:00.454)       0:00:19.746 ******** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set validation result to passed if no test failed] ***********************\nSaturday 21 March 2026  01:17:33 +0000 (0:00:00.293)       0:00:20.040 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Set validation result to failed if a test failed] ************************\nSaturday 21 March 2026  01:17:33 +0000 (0:00:00.249)       0:00:20.290 ******** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nSaturday 21 March 2026  01:17:34 +0000 (0:00:00.238)       0:00:20.529 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nSaturday 21 March 2026  01:17:35 +0000 (0:00:01.716)       0:00:22.246 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nSaturday 21 March 2026  01:17:36 +0000 (0:00:00.257)       0:00:22.503 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nSaturday 21 March 2026  01:17:36 +0000 (0:00:00.250)       0:00:22.753 ******** \n\nTASK [Flush handlers] **********************************************************\nSaturday 21 March 2026  01:17:36 +0000 (0:00:00.068)       0:00:22.822 ******** \n\nTASK [Flush handlers] **********************************************************\nSaturday 21 March 2026  01:17:36 +0000 (0:00:00.216)       0:00:23.038 ******** \n\nRUNNING HANDLER [Write report file] ********************************************\nSaturday 21 March 2026  01:17:36 +0000 (0:00:00.070)       0:00:23.109 ******** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Print report file information] *******************************************\nSaturday 21 March 2026  01:17:37 +0000 (0:00:01.206)       0:00:24.315 ******** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => {\u001b[0m\n\u001b[0;32m    \"msg\": [\u001b[0m\n\u001b[0;32m        \"Validator run completed.\",\u001b[0m\n\u001b[0;32m        \"You can find the report file here:\",\u001b[0m\n\u001b[0;32m        \"/opt/reports/validator/ceph-osds-validator-2026-03-21T01:17:14+00:00-report.json\",\u001b[0m\n\u001b[0;32m        \"on the following host:\",\u001b[0m\n\u001b[0;32m        \"testbed-manager\"\u001b[0m\n\u001b[0;32m    ]\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=35  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  01:17:38 +0000 (0:00:00.383)       0:00:24.698 ******** \n=============================================================================== \nList ceph LVM volumes and collect data ---------------------------------- 2.51s\nGet ceph osd tree ------------------------------------------------------- 2.01s\nAggregate test results step one ----------------------------------------- 1.72s\nWrite report file ------------------------------------------------------- 1.21s\nGet timestamp for report file ------------------------------------------- 0.95s\nCreate report output directory ------------------------------------------ 0.67s\nGet list of ceph-osd containers on host --------------------------------- 0.67s\nPrint report file information ------------------------------------------- 0.59s\nPrepare test data ------------------------------------------------------- 0.49s\nSet test result to failed if an OSD is not running ---------------------- 0.48s\nFail if count of encrypted OSDs does not match -------------------------- 0.47s\nGet CRUSH node data of each OSD host and root node childs --------------- 0.46s\nFail test if any sub test failed ---------------------------------------- 0.45s\nCalculate OSD devices for each host ------------------------------------- 0.45s\nGet unencrypted and encrypted OSDs -------------------------------------- 0.45s\nPrepare test data ------------------------------------------------------- 0.44s\nPass if count of unencrypted OSDs equals count of OSDs ------------------ 0.44s\nPrepare test data ------------------------------------------------------- 0.42s\nPrint report file information ------------------------------------------- 0.38s\nCalculate OSD devices for each host ------------------------------------- 0.36s\n+ sh -c /opt/configuration/scripts/check/200-infrastructure.sh\n+ set -e\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=true\n++ TEMPEST=true\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ [[ -e /etc/redhat-release ]]\n+ source /etc/os-release\n++ PRETTY_NAME='Ubuntu 24.04.4 LTS'\n++ NAME=Ubuntu\n++ VERSION_ID=24.04\n++ VERSION='24.04.4 LTS (Noble Numbat)'\n++ VERSION_CODENAME=noble\n++ ID=ubuntu\n++ ID_LIKE=debian\n++ HOME_URL=https://www.ubuntu.com/\n++ SUPPORT_URL=https://help.ubuntu.com/\n++ BUG_REPORT_URL=https://bugs.launchpad.net/ubuntu/\n++ PRIVACY_POLICY_URL=https://www.ubuntu.com/legal/terms-and-policies/privacy-policy\n++ UBUNTU_CODENAME=noble\n++ LOGO=ubuntu-logo\n+ [[ ubuntu == \\u\\b\\u\\n\\t\\u ]]\n+ packages='libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client'\n+ dpkg -s libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client\n+ sudo apt-get install -y libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client\n\n# Status of Elasticsearch\n\n+ pushd /opt/configuration/contrib\n+ echo\n+ echo '# Status of Elasticsearch'\n+ echo\n+ bash nagios-plugins/check_elasticsearch -H api-int.testbed.osism.xyz -s\nOK - elasticsearch (kolla_logging) is running. status: green; timed_out: false; number_of_nodes: 3; number_of_data_nodes: 3; active_primary_shards: 9; active_shards: 22; relocating_shards: 0; initializing_shards: 0; delayed_unassigned_shards: 0; unassigned_shards: 0  | 'active_primary'=9 'active'=22 'relocating'=0 'init'=0 'delay_unass'=0 'unass'=0\n\n# Status of MariaDB\n\n+ echo\n+ echo '# Status of MariaDB'\n+ echo\n++ semver latest 10.0.0-0\n+ [[ -1 -ge 0 ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ osism status database\n\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConnecting to MariaDB at 192.168.16.9 as root_shard_0...\u001b[0m\n\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster Status: Primary\u001b[0m\n\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConnected: ON\u001b[0m\n\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mReady: ON\u001b[0m\n\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster Size: 3\u001b[0m\n\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLocal State: Synced\u001b[0m\n\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster State UUID: 7e30386f-24c0-11f1-a9de-baebd475a8d6\u001b[0m\n\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster Members: 192.168.16.11:3306,192.168.16.12:3306,192.168.16.10:3306\u001b[0m\n\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGalera Version: 26.4.25(r7387a566)\u001b[0m\n\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLocal Node UUID: b04005f7-24c0-11f1-9f1a-2e2443d031b0\u001b[0m\n\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlow Control Paused: 0.00%\u001b[0m\n\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRecv Queue Avg: 0.00884956\u001b[0m\n\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSend Queue Avg: 0.000591366\u001b[0m\n\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTransactions: 4520 local commits, 6705 replicated, 113 received\u001b[0m\n\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConflicts: 0 cert failures, 0 bf aborts\u001b[0m\n\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMariaDB Uptime: 22 minutes, 21 seconds\u001b[0m\n\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mThreads: 126 connected, 1 running\u001b[0m\n\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mQueries: 237598 total, 0 slow\u001b[0m\n\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAborted Connects: 120\u001b[0m\n\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMariaDB Galera Cluster validation PASSED\u001b[0m\n\n# Status of Prometheus\n\n+ echo\n+ echo '# Status of Prometheus'\n+ echo\n+ curl -s https://api-int.testbed.osism.xyz:9091/-/healthy\nUnauthorized\n+ curl -s https://api-int.testbed.osism.xyz:9091/-/ready\nUnauthorized\n\n# Status of RabbitMQ\n\n+ echo\n+ echo '# Status of RabbitMQ'\n+ echo\n++ semver latest 10.0.0-0\n+ [[ -1 -ge 0 ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ osism status messaging\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Connecting to RabbitMQ Management API at 192.168.16.10:15672 as openstack...\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] RabbitMQ Version: 3.13.7\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Erlang Version: 26.2.5.15\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Cluster Name: rabbit@testbed-node-0\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Cluster Size: 3\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Running Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Partitions: None (healthy)\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Connections: 205, Channels: 204, Queues: 173\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Messages: 218 total, 218 ready, 0 unacked\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Message Rates: 7.8/s publish, 8.2/s deliver\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Disk Free: 58.1 GB (limit: 0.0 GB)\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Memory Used: 0.18 GB (limit: 12.54 GB)\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] File Descriptors: 116/1024\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Sockets: 68/832\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Connecting to RabbitMQ Management API at 192.168.16.11:15672 as openstack...\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] RabbitMQ Version: 3.13.7\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Erlang Version: 26.2.5.15\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Cluster Name: rabbit@testbed-node-1\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Cluster Size: 3\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Running Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Partitions: None (healthy)\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Connections: 205, Channels: 204, Queues: 173\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Messages: 218 total, 218 ready, 0 unacked\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Message Rates: 7.8/s publish, 8.2/s deliver\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Disk Free: 58.4 GB (limit: 0.0 GB)\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Memory Used: 0.18 GB (limit: 12.54 GB)\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] File Descriptors: 124/1024\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Sockets: 78/832\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Connecting to RabbitMQ Management API at 192.168.16.12:15672 as openstack...\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] RabbitMQ Version: 3.13.7\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Erlang Version: 26.2.5.15\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Cluster Name: rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Cluster Size: 3\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Running Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Partitions: None (healthy)\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Connections: 205, Channels: 204, Queues: 173\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Messages: 218 total, 218 ready, 0 unacked\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Message Rates: 7.8/s publish, 8.2/s deliver\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Disk Free: 58.4 GB (limit: 0.0 GB)\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Memory Used: 0.17 GB (limit: 12.54 GB)\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] File Descriptors: 105/1024\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Sockets: 59/832\u001b[0m\n\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRabbitMQ Cluster validation PASSED\u001b[0m\n\n# Status of Redis\n\n+ echo\n+ echo '# Status of Redis'\n+ echo\n+ /usr/lib/nagios/plugins/check_tcp -H 192.168.16.10 -p 6379 -A -E -s 'AUTH QHNA1SZRlOKzLADhUd5ZDgpHfQe6dNfr3bwEdY24\\r\\nPING\\r\\nINFO replication\\r\\nQUIT\\r\\n' -e PONG -e role:master -e slave0:ip=192.168.16.1 -e,port=6379 -j\nTCP OK - 0.001 second response time on 192.168.16.10 port 6379|time=0.001083s;;;0.000000;10.000000\n\n+ popd\n+ echo\n# Create backup of MariaDB database\n+ echo '# Create backup of MariaDB database'\n+ echo\n\n+ osism apply mariadb_backup -e mariadb_backup_type=full\n\u001b[32m2026-03-21 01:18:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of mariadb_backup.\u001b[0m\n\u001b[32m2026-03-21 01:18:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09f29ddb-90dc-4859-b8a6-c08db303db76 (mariadb_backup) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 01:18:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 09f29ddb-90dc-4859-b8a6-c08db303db76 (mariadb_backup) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nSaturday 21 March 2026  01:18:16 +0000 (0:00:00.237)       0:00:00.237 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nSaturday 21 March 2026  01:18:16 +0000 (0:00:00.299)       0:00:00.536 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_mariadb_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_mariadb_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_mariadb_True)\u001b[0m\n\nPLAY [Apply role mariadb] ******************************************************\n\nTASK [mariadb : Group MariaDB hosts based on shards] ***************************\nSaturday 21 March 2026  01:18:17 +0000 (0:00:00.407)       0:00:00.944 ******** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nSaturday 21 March 2026  01:18:17 +0000 (0:00:00.402)       0:00:01.347 ******** \n\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/backup.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [mariadb : Get MariaDB container facts] ***********************************\nSaturday 21 March 2026  01:18:18 +0000 (0:00:00.623)       0:00:01.971 ******** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Taking full database backup via Mariabackup] *******************\nSaturday 21 March 2026  01:18:21 +0000 (0:00:03.368)       0:00:05.340 ******** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_restart\u001b[0m\n\nPLAY [Restart mariadb services] ************************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_start\u001b[0m\n\nPLAY [Start mariadb services] **************************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m\n\u001b[1;35mmariadb_bootstrap_restart\u001b[0m\n\nPLAY [Restart bootstrap mariadb service] ***************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Apply mariadb post-configuration] ****************************************\n\nTASK [Include mariadb post-deploy.yml] *****************************************\nSaturday 21 March 2026  01:20:00 +0000 (0:01:39.326)       0:01:44.667 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Include mariadb post-upgrade.yml] ****************************************\nSaturday 21 March 2026  01:20:01 +0000 (0:00:00.277)       0:01:44.944 ******** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=6   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=4   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=4   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  01:20:01 +0000 (0:00:00.212)       0:01:45.157 ******** \n=============================================================================== \nmariadb : Taking full database backup via Mariabackup ------------------ 99.33s\nmariadb : Get MariaDB container facts ----------------------------------- 3.37s\nmariadb : include_tasks ------------------------------------------------- 0.62s\nGroup hosts based on enabled services ----------------------------------- 0.41s\nmariadb : Group MariaDB hosts based on shards --------------------------- 0.40s\nGroup hosts based on Kolla action --------------------------------------- 0.30s\nInclude mariadb post-deploy.yml ----------------------------------------- 0.28s\nInclude mariadb post-upgrade.yml ---------------------------------------- 0.21s\n+ sh -c /opt/configuration/scripts/check/300-openstack.sh\n+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/configuration/scripts/manager-version.sh\n+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml\n\n# OpenStack endpoints\n\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n+ export OS_CLOUD=admin\n+ OS_CLOUD=admin\n+ echo\n+ echo '# OpenStack endpoints'\n+ echo\n+ openstack endpoint list\n+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+\n| ID                               | Region    | Service Name | Service Type    | Enabled | Interface | URL                                                                 |\n+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+\n| 045abbc9413e49eab3d35b5b46ac273d | RegionOne | octavia      | load-balancer   | True    | internal  | https://api-int.testbed.osism.xyz:9876                              |\n| 1d06f6cc268149419c413b9ef1f0325f | RegionOne | barbican     | key-manager     | True    | internal  | https://api-int.testbed.osism.xyz:9311                              |\n| 205ecdb1bc334998aac9bdfbb6df373a | RegionOne | nova         | compute         | True    | internal  | https://api-int.testbed.osism.xyz:8774/v2.1                         |\n| 48009575ba874f9881df4ddc8a9306f9 | RegionOne | swift        | object-store    | True    | internal  | https://api-int.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s |\n| 48e1442c694a4a8fa468b7430f3c0f24 | RegionOne | barbican     | key-manager     | True    | public    | https://api.testbed.osism.xyz:9311                                  |\n| 577becedbb42491e82fb5b2dea82a3ee | RegionOne | glance       | image           | True    | public    | https://api.testbed.osism.xyz:9292                                  |\n| 57daba2f887740ba8be672a34321ba57 | RegionOne | placement    | placement       | True    | public    | https://api.testbed.osism.xyz:8780                                  |\n| 59fb3fb105524391bae1670c0e48d706 | RegionOne | magnum       | container-infra | True    | internal  | https://api-int.testbed.osism.xyz:9511/v1                           |\n| 6ba274fecd534959bf94f16b1484f3ed | RegionOne | cinderv3     | volumev3        | True    | public    | https://api.testbed.osism.xyz:8776/v3/%(tenant_id)s                 |\n| 84ff3c7b93ca4a4799e5406c09d29dbf | RegionOne | placement    | placement       | True    | internal  | https://api-int.testbed.osism.xyz:8780                              |\n| 86304183b84b4ebdbca0cb27fc1aa03c | RegionOne | magnum       | container-infra | True    | public    | https://api.testbed.osism.xyz:9511/v1                               |\n| b060be793e0c48a5a67a8a574e893250 | RegionOne | designate    | dns             | True    | public    | https://api.testbed.osism.xyz:9001                                  |\n| bb106454a2a9409f9bd8ce1b5864e0ef | RegionOne | designate    | dns             | True    | internal  | https://api-int.testbed.osism.xyz:9001                              |\n| bece2b03d3824cd1b7f3e45a5a595172 | RegionOne | glance       | image           | True    | internal  | https://api-int.testbed.osism.xyz:9292                              |\n| bf1616dcfb4647999288d99d32b6f9c6 | RegionOne | keystone     | identity        | True    | internal  | https://api-int.testbed.osism.xyz:5000                              |\n| db9674b056614c71a5ed51993fbb258e | RegionOne | neutron      | network         | True    | internal  | https://api-int.testbed.osism.xyz:9696                              |\n| e0e8429d47034e81be4c454687ce067b | RegionOne | swift        | object-store    | True    | public    | https://api.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s     |\n| e0f40e46c018498d99911bcbc5191bf3 | RegionOne | keystone     | identity        | True    | public    | https://api.testbed.osism.xyz:5000                                  |\n| e3a16ff8f02c4c3faa07372cf5be5f4e | RegionOne | octavia      | load-balancer   | True    | public    | https://api.testbed.osism.xyz:9876                                  |\n| e6534e6bf126456da312787bee2f2088 | RegionOne | nova         | compute         | True    | public    | https://api.testbed.osism.xyz:8774/v2.1                             |\n| eba1df94bb6d45638c30071431b6fdee | RegionOne | cinderv3     | volumev3        | True    | internal  | https://api-int.testbed.osism.xyz:8776/v3/%(tenant_id)s             |\n| ece1fe5797a742e09c884f898f3f4451 | RegionOne | neutron      | network         | True    | public    | https://api.testbed.osism.xyz:9696                                  |\n+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+\n\n# Cinder\n\n+ echo\n+ echo '# Cinder'\n+ echo\n+ openstack volume service list\n+------------------+----------------------------+----------+---------+-------+----------------------------+\n| Binary           | Host                       | Zone     | Status  | State | Updated At                 |\n+------------------+----------------------------+----------+---------+-------+----------------------------+\n| cinder-scheduler | testbed-node-0             | internal | enabled | up    | 2026-03-21T01:20:02.000000 |\n| cinder-scheduler | testbed-node-2             | internal | enabled | up    | 2026-03-21T01:20:01.000000 |\n| cinder-scheduler | testbed-node-1             | internal | enabled | up    | 2026-03-21T01:20:02.000000 |\n| cinder-volume    | testbed-node-0@rbd-volumes | nova     | enabled | up    | 2026-03-21T01:20:01.000000 |\n| cinder-volume    | testbed-node-1@rbd-volumes | nova     | enabled | up    | 2026-03-21T01:20:03.000000 |\n| cinder-volume    | testbed-node-2@rbd-volumes | nova     | enabled | up    | 2026-03-21T01:20:04.000000 |\n| cinder-backup    | testbed-node-0             | nova     | enabled | up    | 2026-03-21T01:20:07.000000 |\n| cinder-backup    | testbed-node-2             | nova     | enabled | up    | 2026-03-21T01:19:58.000000 |\n| cinder-backup    | testbed-node-1             | nova     | enabled | up    | 2026-03-21T01:19:58.000000 |\n+------------------+----------------------------+----------+---------+-------+----------------------------+\n\n# Neutron\n\n+ echo\n+ echo '# Neutron'\n+ echo\n+ openstack network agent list\n+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+\n| ID                                   | Agent Type                   | Host           | Availability Zone | Alive | State | Binary                     |\n+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+\n| testbed-node-2                       | OVN Controller Gateway agent | testbed-node-2 | nova              | :-)   | UP    | ovn-controller             |\n| testbed-node-5                       | OVN Controller agent         | testbed-node-5 |                   | :-)   | UP    | ovn-controller             |\n| testbed-node-3                       | OVN Controller agent         | testbed-node-3 |                   | :-)   | UP    | ovn-controller             |\n| testbed-node-0                       | OVN Controller Gateway agent | testbed-node-0 | nova              | :-)   | UP    | ovn-controller             |\n| testbed-node-4                       | OVN Controller agent         | testbed-node-4 |                   | :-)   | UP    | ovn-controller             |\n| testbed-node-1                       | OVN Controller Gateway agent | testbed-node-1 | nova              | :-)   | UP    | ovn-controller             |\n| e645415a-98f5-5758-8cd1-c47af282b5c0 | OVN Metadata agent           | testbed-node-3 |                   | :-)   | UP    | neutron-ovn-metadata-agent |\n| 36b9d21c-9928-5c0a-9b27-73ac7a3e770c | OVN Metadata agent           | testbed-node-5 |                   | :-)   | UP    | neutron-ovn-metadata-agent |\n| 4939696e-6092-5a33-bb73-b850064684df | OVN Metadata agent           | testbed-node-4 |                   | :-)   | UP    | neutron-ovn-metadata-agent |\n+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+\n+ openstack network service provider list\n+---------------+------+---------+\n| Service Type  | Name | Default |\n+---------------+------+---------+\n| L3_ROUTER_NAT | ovn  | True    |\n+---------------+------+---------+\n\n+ echo\n+ echo '# Nova'\n# Nova\n+ echo\n\n+ openstack compute service list\n+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+\n| ID                                   | Binary         | Host           | Zone     | Status  | State | Updated At                 |\n+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+\n| 2dfd5f0f-b0d5-4e16-b497-90455e00daac | nova-scheduler | testbed-node-0 | internal | enabled | up    | 2026-03-21T01:20:15.000000 |\n| 66935adb-cd12-41ae-9556-8fe2ef82f37d | nova-scheduler | testbed-node-2 | internal | enabled | up    | 2026-03-21T01:20:16.000000 |\n| bbc1bc14-a891-4a3b-95ca-4222f02fc912 | nova-scheduler | testbed-node-1 | internal | enabled | up    | 2026-03-21T01:20:06.000000 |\n| 62c71ef9-898e-4407-85e2-b6a57bf4b607 | nova-conductor | testbed-node-0 | internal | enabled | up    | 2026-03-21T01:20:08.000000 |\n| ecb7670e-6157-4994-b001-5d6e1cd0b540 | nova-conductor | testbed-node-2 | internal | enabled | up    | 2026-03-21T01:20:09.000000 |\n| 40f2b6fa-67a1-4318-8e28-dec815488cb1 | nova-conductor | testbed-node-1 | internal | enabled | up    | 2026-03-21T01:20:09.000000 |\n| 31494761-ad41-4745-9b1c-34b1a54d475e | nova-compute   | testbed-node-3 | nova     | enabled | up    | 2026-03-21T01:20:07.000000 |\n| 692abb40-5d07-4597-8392-29ef6c60842a | nova-compute   | testbed-node-4 | nova     | enabled | up    | 2026-03-21T01:20:08.000000 |\n| 06ef3854-d0a7-429b-9792-9aa3169e2ca2 | nova-compute   | testbed-node-5 | nova     | enabled | up    | 2026-03-21T01:20:07.000000 |\n+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+\n+ openstack hypervisor list\n+--------------------------------------+---------------------+-----------------+---------------+-------+\n| ID                                   | Hypervisor Hostname | Hypervisor Type | Host IP       | State |\n+--------------------------------------+---------------------+-----------------+---------------+-------+\n| 49c11798-5996-4b3e-a240-8a1cf8c1b2f5 | testbed-node-3      | QEMU            | 192.168.16.13 | up    |\n| 856225ce-2980-4f96-85fa-cef245de4a09 | testbed-node-5      | QEMU            | 192.168.16.15 | up    |\n| b98edce9-70d1-426b-a3d6-43717913d625 | testbed-node-4      | QEMU            | 192.168.16.14 | up    |\n+--------------------------------------+---------------------+-----------------+---------------+-------+\n\n# Run OpenStack test play\n\n+ echo\n+ echo '# Run OpenStack test play'\n+ echo\n+ osism apply --environment openstack test\n\u001b[32m2026-03-21 01:20:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play test in environment openstack\u001b[0m\n\u001b[32m2026-03-21 01:20:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of test.\u001b[0m\n\u001b[32m2026-03-21 01:20:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c177e24-c00d-4262-a7a1-a17b71dc28f9 (test) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 01:20:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 0c177e24-c00d-4262-a7a1-a17b71dc28f9 (test) has been started and output is visible here.\u001b[0m\n\nPLAY [Create test project] *****************************************************\n\nTASK [Create test domain] ******************************************************\nSaturday 21 March 2026  01:20:25 +0000 (0:00:00.103)       0:00:00.103 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test-admin user] **************************************************\nSaturday 21 March 2026  01:20:28 +0000 (0:00:03.558)       0:00:03.662 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Add manager role to user test-admin] *************************************\nSaturday 21 March 2026  01:20:32 +0000 (0:00:03.828)       0:00:07.491 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test project] *****************************************************\nSaturday 21 March 2026  01:20:38 +0000 (0:00:06.134)       0:00:13.625 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test user] ********************************************************\nSaturday 21 March 2026  01:20:42 +0000 (0:00:04.213)       0:00:17.838 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Add member roles to user test] *******************************************\nSaturday 21 March 2026  01:20:46 +0000 (0:00:04.111)       0:00:21.950 ******** \n\u001b[0;33mchanged: [localhost] => (item=load-balancer_member)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=member)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=creator)\u001b[0m\n\nTASK [Create test server group] ************************************************\nSaturday 21 March 2026  01:20:58 +0000 (0:00:11.821)       0:00:33.771 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create ssh security group] ***********************************************\nSaturday 21 March 2026  01:21:03 +0000 (0:00:04.407)       0:00:38.178 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Add rule to ssh security group] ******************************************\nSaturday 21 March 2026  01:21:07 +0000 (0:00:04.579)       0:00:42.758 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create icmp security group] **********************************************\nSaturday 21 March 2026  01:21:11 +0000 (0:00:04.120)       0:00:46.878 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Add rule to icmp security group] *****************************************\nSaturday 21 March 2026  01:21:15 +0000 (0:00:04.101)       0:00:50.980 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test keypair] *****************************************************\nSaturday 21 March 2026  01:21:20 +0000 (0:00:04.105)       0:00:55.085 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test network] *****************************************************\nSaturday 21 March 2026  01:21:23 +0000 (0:00:03.968)       0:00:59.054 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test subnet] ******************************************************\nSaturday 21 March 2026  01:21:29 +0000 (0:00:05.281)       0:01:04.335 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test router] ******************************************************\nSaturday 21 March 2026  01:21:35 +0000 (0:00:05.971)       0:01:10.307 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nPLAY [Manage test instances and volumes] ***************************************\n\nTASK [Get test server group] ***************************************************\nSaturday 21 March 2026  01:21:45 +0000 (0:00:10.595)       0:01:20.902 ******** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nTASK [Detach test volume] ******************************************************\nSaturday 21 March 2026  01:21:49 +0000 (0:00:03.387)       0:01:24.289 ******** \n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nTASK [Delete test volume] ******************************************************\nSaturday 21 March 2026  01:21:49 +0000 (0:00:00.057)       0:01:24.347 ******** \n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nTASK [Delete test instances] ***************************************************\nSaturday 21 March 2026  01:21:49 +0000 (0:00:00.075)       0:01:24.422 ******** \n\u001b[0;36mskipping: [localhost] => (item=test-4) \u001b[0m\n\u001b[0;36mskipping: [localhost] => (item=test-3) \u001b[0m\n\u001b[0;36mskipping: [localhost] => (item=test-2) \u001b[0m\n\u001b[0;36mskipping: [localhost] => (item=test-1) \u001b[0m\n\u001b[0;36mskipping: [localhost] => (item=test) \u001b[0m\n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nTASK [Wait for instance deletion to complete] **********************************\nSaturday 21 March 2026  01:21:49 +0000 (0:00:00.143)       0:01:24.565 ******** \n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nTASK [Create test instances] ***************************************************\nSaturday 21 March 2026  01:21:49 +0000 (0:00:00.124)       0:01:24.690 ******** \n\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m\n\nTASK [Wait for instance creation to complete] **********************************\nSaturday 21 March 2026  01:21:53 +0000 (0:00:04.086)       0:01:28.776 ******** \n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (60 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (59 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (58 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (57 retries left).\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j114091529227.2736', 'results_file': '/ansible/.ansible_async/j114091529227.2736', 'changed': True, 'item': 'test', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j550367883200.2761', 'results_file': '/ansible/.ansible_async/j550367883200.2761', 'changed': True, 'item': 'test-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j292689331656.2786', 'results_file': '/ansible/.ansible_async/j292689331656.2786', 'changed': True, 'item': 'test-2', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j923253320887.2818', 'results_file': '/ansible/.ansible_async/j923253320887.2818', 'changed': True, 'item': 'test-3', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j335587441571.2843', 'results_file': '/ansible/.ansible_async/j335587441571.2843', 'changed': True, 'item': 'test-4', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [Add metadata to instances] ***********************************************\nSaturday 21 March 2026  01:22:40 +0000 (0:00:47.008)       0:02:15.785 ******** \n\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m\n\nTASK [Wait for metadata to be added] *******************************************\nSaturday 21 March 2026  01:22:45 +0000 (0:00:04.580)       0:02:20.365 ******** \n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for metadata to be added (30 retries left).\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j529373546002.2940', 'results_file': '/ansible/.ansible_async/j529373546002.2940', 'changed': True, 'item': 'test', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j977543273719.2965', 'results_file': '/ansible/.ansible_async/j977543273719.2965', 'changed': True, 'item': 'test-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j868334366829.2990', 'results_file': '/ansible/.ansible_async/j868334366829.2990', 'changed': True, 'item': 'test-2', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j116173683184.3015', 'results_file': '/ansible/.ansible_async/j116173683184.3015', 'changed': True, 'item': 'test-3', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j213789100609.3040', 'results_file': '/ansible/.ansible_async/j213789100609.3040', 'changed': True, 'item': 'test-4', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [Add tag to instances] ****************************************************\nSaturday 21 March 2026  01:22:54 +0000 (0:00:09.169)       0:02:29.535 ******** \n\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m\n\nTASK [Wait for tags to be added] ***********************************************\nSaturday 21 March 2026  01:22:58 +0000 (0:00:04.306)       0:02:33.842 ******** \n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for tags to be added (30 retries left).\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j164432034846.3116', 'results_file': '/ansible/.ansible_async/j164432034846.3116', 'changed': True, 'item': 'test', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j544760755925.3141', 'results_file': '/ansible/.ansible_async/j544760755925.3141', 'changed': True, 'item': 'test-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j529594319159.3167', 'results_file': '/ansible/.ansible_async/j529594319159.3167', 'changed': True, 'item': 'test-2', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j894484329495.3193', 'results_file': '/ansible/.ansible_async/j894484329495.3193', 'changed': True, 'item': 'test-3', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j226104707254.3219', 'results_file': '/ansible/.ansible_async/j226104707254.3219', 'changed': True, 'item': 'test-4', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [Create test volume] ******************************************************\nSaturday 21 March 2026  01:23:08 +0000 (0:00:09.356)       0:02:43.199 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Attach test volume] ******************************************************\nSaturday 21 March 2026  01:23:14 +0000 (0:00:06.545)       0:02:49.745 ******** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create floating ip address] **********************************************\nSaturday 21 March 2026  01:23:28 +0000 (0:00:13.448)       0:03:03.193 ******** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nTASK [Print floating ip address] ***********************************************\nSaturday 21 March 2026  01:23:32 +0000 (0:00:04.489)       0:03:07.683 ******** \n\u001b[0;32mok: [localhost] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"192.168.112.150\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  01:23:32 +0000 (0:00:00.040)       0:03:07.724 ******** \n=============================================================================== \nWait for instance creation to complete --------------------------------- 47.01s\nAttach test volume ----------------------------------------------------- 13.45s\nAdd member roles to user test ------------------------------------------ 11.82s\nCreate test router ----------------------------------------------------- 10.60s\nWait for tags to be added ----------------------------------------------- 9.36s\nWait for metadata to be added ------------------------------------------- 9.17s\nCreate test volume ------------------------------------------------------ 6.55s\nAdd manager role to user test-admin ------------------------------------- 6.13s\nCreate test subnet ------------------------------------------------------ 5.97s\nCreate test network ----------------------------------------------------- 5.28s\nAdd metadata to instances ----------------------------------------------- 4.58s\nCreate ssh security group ----------------------------------------------- 4.58s\nCreate floating ip address ---------------------------------------------- 4.49s\nCreate test server group ------------------------------------------------ 4.41s\nAdd tag to instances ---------------------------------------------------- 4.31s\nCreate test project ----------------------------------------------------- 4.21s\nAdd rule to ssh security group ------------------------------------------ 4.12s\nCreate test user -------------------------------------------------------- 4.11s\nAdd rule to icmp security group ----------------------------------------- 4.11s\nCreate icmp security group ---------------------------------------------- 4.10s\n+ server_list\n+ openstack --os-cloud test server list\n+--------------------------------------+--------+--------+---------------------------------------+--------------------------+----------+\n| ID                                   | Name   | Status | Networks                              | Image                    | Flavor   |\n+--------------------------------------+--------+--------+---------------------------------------+--------------------------+----------+\n| e341f1ff-bd52-468b-8137-05723da11652 | test-4 | ACTIVE | test=192.168.112.145, 192.168.200.240 | N/A (booted from volume) | SCS-1L-1 |\n| a219ecb7-4093-44bb-b939-a453113dfec2 | test-3 | ACTIVE | test=192.168.112.103, 192.168.200.23  | N/A (booted from volume) | SCS-1L-1 |\n| ef7382d2-8009-467e-b354-44b46a384980 | test-2 | ACTIVE | test=192.168.112.117, 192.168.200.243 | N/A (booted from volume) | SCS-1L-1 |\n| 4740b0e6-2766-4b54-a663-63e0e558e778 | test   | ACTIVE | test=192.168.112.150, 192.168.200.203 | N/A (booted from volume) | SCS-1L-1 |\n| 56e68f34-0e54-4c15-a73d-f764465f4df9 | test-1 | ACTIVE | test=192.168.112.200, 192.168.200.144 | N/A (booted from volume) | SCS-1L-1 |\n+--------------------------------------+--------+--------+---------------------------------------+--------------------------+----------+\n+ openstack --os-cloud test server show test\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-03-21T01:22:24.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.150, 192.168.200.203                                                                                                                                                                                                                                                                                                                                               |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-03-21T01:21:58Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='True', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | 7db2ef38c4e8041c1ab68e4105682d7049680cf0587156722412a340                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | 4740b0e6-2766-4b54-a663-63e0e558e778                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | ce936a29388e4c2e9c53a72b276e2fb2                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test'                                                                                                                                                                                                                                                                                                                                                                     |\n| security_groups                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n|                                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-03-21T01:22:46Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | 413e209667b94f5e9151b49aab438337                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='383f89d6-5820-4fa6-9c5d-f29086bb0799'                                                                                                                                                                                                                                                                                                             |\n|                                     | delete_on_termination='False', id='96ecdebe-a1b7-4c81-99fd-d6c165f8a05d'                                                                                                                                                                                                                                                                                                            |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ openstack --os-cloud test server show test-1\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test-1                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-03-21T01:22:24.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.200, 192.168.200.144                                                                                                                                                                                                                                                                                                                                               |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-03-21T01:21:58Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='True', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | 7db2ef38c4e8041c1ab68e4105682d7049680cf0587156722412a340                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | 56e68f34-0e54-4c15-a73d-f764465f4df9                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test-1                                                                                                                                                                                                                                                                                                                                                                              |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | ce936a29388e4c2e9c53a72b276e2fb2                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test-1'                                                                                                                                                                                                                                                                                                                                                                   |\n| security_groups                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n|                                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-03-21T01:22:46Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | 413e209667b94f5e9151b49aab438337                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='bedb1497-49f1-4ba2-8d84-4c80dfa714cf'                                                                                                                                                                                                                                                                                                             |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ openstack --os-cloud test server show test-2\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test-2                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-03-21T01:22:26.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.117, 192.168.200.243                                                                                                                                                                                                                                                                                                                                               |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-03-21T01:21:59Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='True', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | 6ede240e48d06e40d4d6e3ffd457e1e0ed1e43d51a67d24c33c067cf                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | ef7382d2-8009-467e-b354-44b46a384980                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test-2                                                                                                                                                                                                                                                                                                                                                                              |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | ce936a29388e4c2e9c53a72b276e2fb2                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test-2'                                                                                                                                                                                                                                                                                                                                                                   |\n| security_groups                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n|                                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-03-21T01:22:47Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | 413e209667b94f5e9151b49aab438337                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='39b5d04c-efff-4103-8694-5044b75d8ed3'                                                                                                                                                                                                                                                                                                             |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ openstack --os-cloud test server show test-3\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test-3                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-03-21T01:22:26.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.103, 192.168.200.23                                                                                                                                                                                                                                                                                                                                                |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-03-21T01:21:59Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='True', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | 6ede240e48d06e40d4d6e3ffd457e1e0ed1e43d51a67d24c33c067cf                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | a219ecb7-4093-44bb-b939-a453113dfec2                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test-3                                                                                                                                                                                                                                                                                                                                                                              |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | ce936a29388e4c2e9c53a72b276e2fb2                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test-3'                                                                                                                                                                                                                                                                                                                                                                   |\n| security_groups                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n|                                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-03-21T01:22:48Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | 413e209667b94f5e9151b49aab438337                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='bf295bac-2ea3-4452-914f-011aaf87a95b'                                                                                                                                                                                                                                                                                                             |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ openstack --os-cloud test server show test-4\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test-4                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-03-21T01:22:25.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.145, 192.168.200.240                                                                                                                                                                                                                                                                                                                                               |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-03-21T01:22:01Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='True', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | 6ede240e48d06e40d4d6e3ffd457e1e0ed1e43d51a67d24c33c067cf                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | e341f1ff-bd52-468b-8137-05723da11652                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test-4                                                                                                                                                                                                                                                                                                                                                                              |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | ce936a29388e4c2e9c53a72b276e2fb2                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test-4'                                                                                                                                                                                                                                                                                                                                                                   |\n| security_groups                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n|                                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-03-21T01:22:49Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | 413e209667b94f5e9151b49aab438337                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='c3f0bedd-88a5-41cd-992f-4319ff75bf1e'                                                                                                                                                                                                                                                                                                             |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ server_ping\n++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'\n++ tr -d '\\r'\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.103\nPING 192.168.112.103 (192.168.112.103) 56(84) bytes of data.\n64 bytes from 192.168.112.103: icmp_seq=1 ttl=63 time=3.68 ms\n64 bytes from 192.168.112.103: icmp_seq=2 ttl=63 time=1.68 ms\n64 bytes from 192.168.112.103: icmp_seq=3 ttl=63 time=1.24 ms\n\n--- 192.168.112.103 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2004ms\nrtt min/avg/max/mdev = 1.240/2.202/3.684/1.063 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.117\nPING 192.168.112.117 (192.168.112.117) 56(84) bytes of data.\n64 bytes from 192.168.112.117: icmp_seq=1 ttl=63 time=4.32 ms\n64 bytes from 192.168.112.117: icmp_seq=2 ttl=63 time=1.38 ms\n64 bytes from 192.168.112.117: icmp_seq=3 ttl=63 time=1.09 ms\n\n--- 192.168.112.117 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2002ms\nrtt min/avg/max/mdev = 1.091/2.262/4.318/1.458 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.150\nPING 192.168.112.150 (192.168.112.150) 56(84) bytes of data.\n64 bytes from 192.168.112.150: icmp_seq=1 ttl=63 time=3.32 ms\n64 bytes from 192.168.112.150: icmp_seq=2 ttl=63 time=1.29 ms\n64 bytes from 192.168.112.150: icmp_seq=3 ttl=63 time=2.79 ms\n\n--- 192.168.112.150 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2002ms\nrtt min/avg/max/mdev = 1.289/2.467/3.322/0.860 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.145\nPING 192.168.112.145 (192.168.112.145) 56(84) bytes of data.\n64 bytes from 192.168.112.145: icmp_seq=1 ttl=63 time=3.60 ms\n64 bytes from 192.168.112.145: icmp_seq=2 ttl=63 time=1.63 ms\n64 bytes from 192.168.112.145: icmp_seq=3 ttl=63 time=0.967 ms\n\n--- 192.168.112.145 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 0.967/2.063/3.595/1.116 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.200\nPING 192.168.112.200 (192.168.112.200) 56(84) bytes of data.\n64 bytes from 192.168.112.200: icmp_seq=1 ttl=63 time=4.57 ms\n64 bytes from 192.168.112.200: icmp_seq=2 ttl=63 time=1.39 ms\n64 bytes from 192.168.112.200: icmp_seq=3 ttl=63 time=1.33 ms\n\n--- 192.168.112.200 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.325/2.427/4.566/1.512 ms\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ compute_list\n+ osism manage compute list testbed-node-3\n\u001b[32m2026-03-21 01:24:05\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-21 01:24:05\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-21 01:24:05\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ osism manage compute list testbed-node-4\n\u001b[32m2026-03-21 01:24:09\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-21 01:24:09\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-21 01:24:09\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| 4740b0e6-2766-4b54-a663-63e0e558e778 | test   | ACTIVE   |\n| 56e68f34-0e54-4c15-a73d-f764465f4df9 | test-1 | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ osism manage compute list testbed-node-5\n\u001b[32m2026-03-21 01:24:13\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-21 01:24:13\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-21 01:24:13\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| e341f1ff-bd52-468b-8137-05723da11652 | test-4 | ACTIVE   |\n| a219ecb7-4093-44bb-b939-a453113dfec2 | test-3 | ACTIVE   |\n| ef7382d2-8009-467e-b354-44b46a384980 | test-2 | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ osism manage compute migrate --yes --target testbed-node-3 testbed-node-4\n\u001b[32m2026-03-21 01:24:16\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-21 01:24:16\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-21 01:24:16\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n\u001b[32m2026-03-21 01:24:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 4740b0e6-2766-4b54-a663-63e0e558e778\u001b[0m\n\u001b[32m2026-03-21 01:24:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:24:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:24:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:24:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:24:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:24:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:24:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:24:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:24:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:24:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:24:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:24:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:25:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-03-21 01:25:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 56e68f34-0e54-4c15-a73d-f764465f4df9\u001b[0m\n\u001b[32m2026-03-21 01:25:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:25:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:25:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:25:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:25:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:25:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:25:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:25:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:25:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:25:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) completed with status ACTIVE\u001b[0m\n+ compute_list\n+ osism manage compute list testbed-node-3\n\u001b[32m2026-03-21 01:25:33\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-21 01:25:33\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-21 01:25:33\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| 4740b0e6-2766-4b54-a663-63e0e558e778 | test   | ACTIVE   |\n| 56e68f34-0e54-4c15-a73d-f764465f4df9 | test-1 | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ osism manage compute list testbed-node-4\n\u001b[32m2026-03-21 01:25:36\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-21 01:25:36\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-21 01:25:36\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ osism manage compute list testbed-node-5\n\u001b[32m2026-03-21 01:25:39\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-21 01:25:39\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-21 01:25:39\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| e341f1ff-bd52-468b-8137-05723da11652 | test-4 | ACTIVE   |\n| a219ecb7-4093-44bb-b939-a453113dfec2 | test-3 | ACTIVE   |\n| ef7382d2-8009-467e-b354-44b46a384980 | test-2 | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ server_ping\n++ tr -d '\\r'\n++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.103\nPING 192.168.112.103 (192.168.112.103) 56(84) bytes of data.\n64 bytes from 192.168.112.103: icmp_seq=1 ttl=63 time=3.75 ms\n64 bytes from 192.168.112.103: icmp_seq=2 ttl=63 time=1.56 ms\n64 bytes from 192.168.112.103: icmp_seq=3 ttl=63 time=1.26 ms\n\n--- 192.168.112.103 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2004ms\nrtt min/avg/max/mdev = 1.260/2.188/3.749/1.110 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.117\nPING 192.168.112.117 (192.168.112.117) 56(84) bytes of data.\n64 bytes from 192.168.112.117: icmp_seq=1 ttl=63 time=5.35 ms\n64 bytes from 192.168.112.117: icmp_seq=2 ttl=63 time=1.28 ms\n64 bytes from 192.168.112.117: icmp_seq=3 ttl=63 time=1.24 ms\n\n--- 192.168.112.117 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.235/2.620/5.348/1.928 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.150\nPING 192.168.112.150 (192.168.112.150) 56(84) bytes of data.\n64 bytes from 192.168.112.150: icmp_seq=1 ttl=63 time=3.55 ms\n64 bytes from 192.168.112.150: icmp_seq=2 ttl=63 time=1.52 ms\n64 bytes from 192.168.112.150: icmp_seq=3 ttl=63 time=1.15 ms\n\n--- 192.168.112.150 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2004ms\nrtt min/avg/max/mdev = 1.152/2.073/3.546/1.052 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.145\nPING 192.168.112.145 (192.168.112.145) 56(84) bytes of data.\n64 bytes from 192.168.112.145: icmp_seq=1 ttl=63 time=3.12 ms\n64 bytes from 192.168.112.145: icmp_seq=2 ttl=63 time=1.37 ms\n64 bytes from 192.168.112.145: icmp_seq=3 ttl=63 time=1.14 ms\n\n--- 192.168.112.145 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.137/1.875/3.121/0.885 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.200\nPING 192.168.112.200 (192.168.112.200) 56(84) bytes of data.\n64 bytes from 192.168.112.200: icmp_seq=1 ttl=63 time=3.30 ms\n64 bytes from 192.168.112.200: icmp_seq=2 ttl=63 time=1.38 ms\n64 bytes from 192.168.112.200: icmp_seq=3 ttl=63 time=1.13 ms\n\n--- 192.168.112.200 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.132/1.934/3.295/0.967 ms\n+ osism manage compute migrate --yes --target testbed-node-3 testbed-node-5\n\u001b[32m2026-03-21 01:25:55\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-21 01:25:55\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-21 01:25:55\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n\u001b[32m2026-03-21 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server e341f1ff-bd52-468b-8137-05723da11652\u001b[0m\n\u001b[32m2026-03-21 01:26:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:26:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:26:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:26:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:26:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:26:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:26:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:26:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:26:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-03-21 01:26:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server a219ecb7-4093-44bb-b939-a453113dfec2\u001b[0m\n\u001b[32m2026-03-21 01:26:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:26:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:26:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:26:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:26:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:26:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:26:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:26:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:26:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-03-21 01:26:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server ef7382d2-8009-467e-b354-44b46a384980\u001b[0m\n\u001b[32m2026-03-21 01:27:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:27:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:27:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:27:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:27:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:27:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:27:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:27:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:27:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) completed with status ACTIVE\u001b[0m\n+ compute_list\n+ osism manage compute list testbed-node-3\n\u001b[32m2026-03-21 01:27:29\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-21 01:27:29\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-21 01:27:29\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| e341f1ff-bd52-468b-8137-05723da11652 | test-4 | ACTIVE   |\n| a219ecb7-4093-44bb-b939-a453113dfec2 | test-3 | ACTIVE   |\n| ef7382d2-8009-467e-b354-44b46a384980 | test-2 | ACTIVE   |\n| 4740b0e6-2766-4b54-a663-63e0e558e778 | test   | ACTIVE   |\n| 56e68f34-0e54-4c15-a73d-f764465f4df9 | test-1 | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ osism manage compute list testbed-node-4\n\u001b[32m2026-03-21 01:27:33\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-21 01:27:33\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-21 01:27:33\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ osism manage compute list testbed-node-5\n\u001b[32m2026-03-21 01:27:35\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-21 01:27:35\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-21 01:27:35\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ server_ping\n++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'\n++ tr -d '\\r'\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.103\nPING 192.168.112.103 (192.168.112.103) 56(84) bytes of data.\n64 bytes from 192.168.112.103: icmp_seq=1 ttl=63 time=3.75 ms\n64 bytes from 192.168.112.103: icmp_seq=2 ttl=63 time=1.62 ms\n64 bytes from 192.168.112.103: icmp_seq=3 ttl=63 time=1.97 ms\n\n--- 192.168.112.103 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.620/2.447/3.754/0.934 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.117\nPING 192.168.112.117 (192.168.112.117) 56(84) bytes of data.\n64 bytes from 192.168.112.117: icmp_seq=1 ttl=63 time=10.7 ms\n64 bytes from 192.168.112.117: icmp_seq=2 ttl=63 time=2.08 ms\n64 bytes from 192.168.112.117: icmp_seq=3 ttl=63 time=1.99 ms\n\n--- 192.168.112.117 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.985/4.911/10.667/4.070 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.150\nPING 192.168.112.150 (192.168.112.150) 56(84) bytes of data.\n64 bytes from 192.168.112.150: icmp_seq=1 ttl=63 time=6.27 ms\n64 bytes from 192.168.112.150: icmp_seq=2 ttl=63 time=2.11 ms\n64 bytes from 192.168.112.150: icmp_seq=3 ttl=63 time=1.55 ms\n\n--- 192.168.112.150 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2002ms\nrtt min/avg/max/mdev = 1.550/3.307/6.267/2.104 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.145\nPING 192.168.112.145 (192.168.112.145) 56(84) bytes of data.\n64 bytes from 192.168.112.145: icmp_seq=1 ttl=63 time=7.33 ms\n64 bytes from 192.168.112.145: icmp_seq=2 ttl=63 time=2.18 ms\n64 bytes from 192.168.112.145: icmp_seq=3 ttl=63 time=1.83 ms\n\n--- 192.168.112.145 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.833/3.779/7.325/2.510 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.200\nPING 192.168.112.200 (192.168.112.200) 56(84) bytes of data.\n64 bytes from 192.168.112.200: icmp_seq=1 ttl=63 time=5.52 ms\n64 bytes from 192.168.112.200: icmp_seq=2 ttl=63 time=1.89 ms\n64 bytes from 192.168.112.200: icmp_seq=3 ttl=63 time=1.42 ms\n\n--- 192.168.112.200 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.419/2.940/5.515/1.830 ms\n+ osism manage compute migrate --yes --target testbed-node-4 testbed-node-3\n\u001b[32m2026-03-21 01:27:51\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-21 01:27:51\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-21 01:27:51\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n\u001b[32m2026-03-21 01:27:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server e341f1ff-bd52-468b-8137-05723da11652\u001b[0m\n\u001b[32m2026-03-21 01:28:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:28:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:28:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:28:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:28:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:28:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:28:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:28:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:28:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-03-21 01:28:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server a219ecb7-4093-44bb-b939-a453113dfec2\u001b[0m\n\u001b[32m2026-03-21 01:28:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:28:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:28:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:28:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:28:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:28:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:28:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:28:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:28:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-03-21 01:28:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server ef7382d2-8009-467e-b354-44b46a384980\u001b[0m\n\u001b[32m2026-03-21 01:29:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:29:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:29:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:29:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:29:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:29:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:29:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:29:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:29:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-03-21 01:29:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 4740b0e6-2766-4b54-a663-63e0e558e778\u001b[0m\n\u001b[32m2026-03-21 01:29:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:29:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:29:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:29:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:29:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:29:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:29:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:29:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:29:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:29:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-03-21 01:29:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 56e68f34-0e54-4c15-a73d-f764465f4df9\u001b[0m\n\u001b[32m2026-03-21 01:30:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:30:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:30:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:30:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:30:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:30:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:30:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:30:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:30:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) completed with status ACTIVE\u001b[0m\n+ compute_list\n+ osism manage compute list testbed-node-3\n\u001b[32m2026-03-21 01:30:25\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-21 01:30:25\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-21 01:30:25\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ osism manage compute list testbed-node-4\n\u001b[32m2026-03-21 01:30:28\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-21 01:30:28\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-21 01:30:28\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| e341f1ff-bd52-468b-8137-05723da11652 | test-4 | ACTIVE   |\n| a219ecb7-4093-44bb-b939-a453113dfec2 | test-3 | ACTIVE   |\n| ef7382d2-8009-467e-b354-44b46a384980 | test-2 | ACTIVE   |\n| 4740b0e6-2766-4b54-a663-63e0e558e778 | test   | ACTIVE   |\n| 56e68f34-0e54-4c15-a73d-f764465f4df9 | test-1 | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ osism manage compute list testbed-node-5\n\u001b[32m2026-03-21 01:30:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-21 01:30:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-21 01:30:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ server_ping\n++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'\n++ tr -d '\\r'\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.103\nPING 192.168.112.103 (192.168.112.103) 56(84) bytes of data.\n64 bytes from 192.168.112.103: icmp_seq=1 ttl=63 time=9.55 ms\n64 bytes from 192.168.112.103: icmp_seq=2 ttl=63 time=2.38 ms\n64 bytes from 192.168.112.103: icmp_seq=3 ttl=63 time=1.70 ms\n\n--- 192.168.112.103 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.701/4.542/9.551/3.552 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.117\nPING 192.168.112.117 (192.168.112.117) 56(84) bytes of data.\n64 bytes from 192.168.112.117: icmp_seq=1 ttl=63 time=7.05 ms\n64 bytes from 192.168.112.117: icmp_seq=2 ttl=63 time=2.38 ms\n64 bytes from 192.168.112.117: icmp_seq=3 ttl=63 time=1.96 ms\n\n--- 192.168.112.117 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.964/3.799/7.054/2.307 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.150\nPING 192.168.112.150 (192.168.112.150) 56(84) bytes of data.\n64 bytes from 192.168.112.150: icmp_seq=1 ttl=63 time=6.13 ms\n64 bytes from 192.168.112.150: icmp_seq=2 ttl=63 time=2.70 ms\n64 bytes from 192.168.112.150: icmp_seq=3 ttl=63 time=1.83 ms\n\n--- 192.168.112.150 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.834/3.554/6.126/1.852 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.145\nPING 192.168.112.145 (192.168.112.145) 56(84) bytes of data.\n64 bytes from 192.168.112.145: icmp_seq=1 ttl=63 time=4.76 ms\n64 bytes from 192.168.112.145: icmp_seq=2 ttl=63 time=2.34 ms\n64 bytes from 192.168.112.145: icmp_seq=3 ttl=63 time=1.90 ms\n\n--- 192.168.112.145 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2004ms\nrtt min/avg/max/mdev = 1.898/2.999/4.761/1.258 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.200\nPING 192.168.112.200 (192.168.112.200) 56(84) bytes of data.\n64 bytes from 192.168.112.200: icmp_seq=1 ttl=63 time=4.50 ms\n64 bytes from 192.168.112.200: icmp_seq=2 ttl=63 time=1.99 ms\n64 bytes from 192.168.112.200: icmp_seq=3 ttl=63 time=1.66 ms\n\n--- 192.168.112.200 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2002ms\nrtt min/avg/max/mdev = 1.655/2.715/4.499/1.268 ms\n+ osism manage compute migrate --yes --target testbed-node-5 testbed-node-4\n\u001b[32m2026-03-21 01:30:48\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-21 01:30:48\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-21 01:30:48\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n\u001b[32m2026-03-21 01:30:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server e341f1ff-bd52-468b-8137-05723da11652\u001b[0m\n\u001b[32m2026-03-21 01:31:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:31:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:31:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:31:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:31:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:31:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:31:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:31:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:31:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-03-21 01:31:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server a219ecb7-4093-44bb-b939-a453113dfec2\u001b[0m\n\u001b[32m2026-03-21 01:31:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:31:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:31:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:31:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:31:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:31:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:31:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:31:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:31:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-03-21 01:31:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server ef7382d2-8009-467e-b354-44b46a384980\u001b[0m\n\u001b[32m2026-03-21 01:31:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:31:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:32:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:32:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:32:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:32:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:32:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:32:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:32:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-03-21 01:32:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 4740b0e6-2766-4b54-a663-63e0e558e778\u001b[0m\n\u001b[32m2026-03-21 01:32:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:32:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:32:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:32:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:32:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:32:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:32:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:32:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:32:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:32:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:32:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-03-21 01:32:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 56e68f34-0e54-4c15-a73d-f764465f4df9\u001b[0m\n\u001b[32m2026-03-21 01:32:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:33:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:33:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:33:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:33:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:33:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:33:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:33:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-21 01:33:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) completed with status ACTIVE\u001b[0m\n+ compute_list\n+ osism manage compute list testbed-node-3\n\u001b[32m2026-03-21 01:33:18\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-21 01:33:18\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-21 01:33:18\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ osism manage compute list testbed-node-4\n\u001b[32m2026-03-21 01:33:21\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-21 01:33:21\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-21 01:33:21\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ osism manage compute list testbed-node-5\n\u001b[32m2026-03-21 01:33:23\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-21 01:33:23\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-21 01:33:23\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| e341f1ff-bd52-468b-8137-05723da11652 | test-4 | ACTIVE   |\n| a219ecb7-4093-44bb-b939-a453113dfec2 | test-3 | ACTIVE   |\n| ef7382d2-8009-467e-b354-44b46a384980 | test-2 | ACTIVE   |\n| 4740b0e6-2766-4b54-a663-63e0e558e778 | test   | ACTIVE   |\n| 56e68f34-0e54-4c15-a73d-f764465f4df9 | test-1 | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ server_ping\n++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'\n++ tr -d '\\r'\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.103\nPING 192.168.112.103 (192.168.112.103) 56(84) bytes of data.\n64 bytes from 192.168.112.103: icmp_seq=1 ttl=63 time=7.19 ms\n64 bytes from 192.168.112.103: icmp_seq=2 ttl=63 time=2.12 ms\n64 bytes from 192.168.112.103: icmp_seq=3 ttl=63 time=1.71 ms\n\n--- 192.168.112.103 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.712/3.672/7.189/2.491 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.117\nPING 192.168.112.117 (192.168.112.117) 56(84) bytes of data.\n64 bytes from 192.168.112.117: icmp_seq=1 ttl=63 time=8.33 ms\n64 bytes from 192.168.112.117: icmp_seq=2 ttl=63 time=1.98 ms\n64 bytes from 192.168.112.117: icmp_seq=3 ttl=63 time=1.75 ms\n\n--- 192.168.112.117 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.752/4.022/8.331/3.048 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.150\nPING 192.168.112.150 (192.168.112.150) 56(84) bytes of data.\n64 bytes from 192.168.112.150: icmp_seq=1 ttl=63 time=4.53 ms\n64 bytes from 192.168.112.150: icmp_seq=2 ttl=63 time=1.60 ms\n64 bytes from 192.168.112.150: icmp_seq=3 ttl=63 time=1.76 ms\n\n--- 192.168.112.150 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2004ms\nrtt min/avg/max/mdev = 1.597/2.629/4.532/1.347 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.145\nPING 192.168.112.145 (192.168.112.145) 56(84) bytes of data.\n64 bytes from 192.168.112.145: icmp_seq=1 ttl=63 time=4.75 ms\n64 bytes from 192.168.112.145: icmp_seq=2 ttl=63 time=2.21 ms\n64 bytes from 192.168.112.145: icmp_seq=3 ttl=63 time=1.82 ms\n\n--- 192.168.112.145 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.820/2.928/4.753/1.300 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.200\nPING 192.168.112.200 (192.168.112.200) 56(84) bytes of data.\n64 bytes from 192.168.112.200: icmp_seq=1 ttl=63 time=4.60 ms\n64 bytes from 192.168.112.200: icmp_seq=2 ttl=63 time=2.18 ms\n64 bytes from 192.168.112.200: icmp_seq=3 ttl=63 time=1.79 ms\n\n--- 192.168.112.200 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.786/2.853/4.599/1.244 ms",
                            "stdout_lines": [
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/configuration/scripts/manager-version.sh",
                                "+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml",
                                "",
                                "# CHECK",
                                "",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "+ echo",
                                "+ echo '# CHECK'",
                                "+ echo",
                                "+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2",
                                "++ semver latest 5.0.0",
                                "",
                                "## Containers @ testbed-manager",
                                "",
                                "+ [[ -1 -eq -1 ]]",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ echo",
                                "+ echo '## Containers @ testbed-manager'",
                                "+ echo",
                                "+ osism container testbed-manager ps",
                                "\u001b[32m2026-03-21 01:15:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCreating empty known_hosts file: /share/known_hosts\u001b[0m",
                                "CONTAINER ID   IMAGE                                                           COMMAND                  CREATED          STATUS                    PORTS                                                                             NAMES",
                                "79dcf2ba75f6   registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2   \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes                                                                                               prometheus_blackbox_exporter",
                                "d0253e1bc915   registry.osism.tech/kolla/prometheus-alertmanager:2024.2        \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                                                                                               prometheus_alertmanager",
                                "debdfb3eb46d   registry.osism.tech/kolla/prometheus-cadvisor:2024.2            \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                                                                                               prometheus_cadvisor",
                                "ae733499c9e4   registry.osism.tech/kolla/prometheus-node-exporter:2024.2       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                                                                                               prometheus_node_exporter",
                                "d34ecdad237a   registry.osism.tech/kolla/prometheus-v2-server:2024.2           \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                                                                                               prometheus_server",
                                "32ffe7d6598e   registry.osism.tech/osism/cephclient:reef                       \"/usr/bin/dumb-init \u2026\"   16 minutes ago   Up 15 minutes                                                                                               cephclient",
                                "b11564e80c95   registry.osism.tech/kolla/cron:2024.2                           \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes                                                                                               cron",
                                "f36f602ac7cc   registry.osism.tech/kolla/kolla-toolbox:2024.2                  \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes                                                                                               kolla_toolbox",
                                "87666c63b6d1   registry.osism.tech/kolla/fluentd:2024.2                        \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                                                                                               fluentd",
                                "3d86d9380bd4   phpmyadmin/phpmyadmin:5.2                                       \"/docker-entrypoint.\u2026\"   29 minutes ago   Up 29 minutes (healthy)   80/tcp                                                                            phpmyadmin",
                                "d2d538d7dc83   registry.osism.tech/osism/openstackclient:2024.2                \"/usr/bin/dumb-init \u2026\"   30 minutes ago   Up 29 minutes                                                                                               openstackclient",
                                "2cd72298f354   registry.osism.tech/osism/homer:v25.10.1                        \"/bin/sh /entrypoint\u2026\"   30 minutes ago   Up 29 minutes (healthy)   8080/tcp                                                                          homer",
                                "f5125536ac8a   registry.osism.tech/dockerhub/ubuntu/squid:6.1-23.10_beta       \"entrypoint.sh -f /e\u2026\"   52 minutes ago   Up 52 minutes (healthy)   192.168.16.5:3128->3128/tcp                                                       squid",
                                "8031abebbadc   registry.osism.tech/osism/inventory-reconciler:latest           \"/sbin/tini -- /entr\u2026\"   57 minutes ago   Up 36 minutes (healthy)                                                                                     manager-inventory_reconciler-1",
                                "98e5144f44c7   registry.osism.tech/osism/osism-ansible:latest                  \"/entrypoint.sh osis\u2026\"   57 minutes ago   Up 36 minutes (healthy)                                                                                     osism-ansible",
                                "198ac8812cab   registry.osism.tech/osism/kolla-ansible:2024.2                  \"/entrypoint.sh osis\u2026\"   57 minutes ago   Up 36 minutes (healthy)                                                                                     kolla-ansible",
                                "ddda498069fb   registry.osism.tech/osism/osism-kubernetes:latest               \"/entrypoint.sh osis\u2026\"   57 minutes ago   Up 36 minutes (healthy)                                                                                     osism-kubernetes",
                                "2a2e2ce2531b   registry.osism.tech/osism/ceph-ansible:reef                     \"/entrypoint.sh osis\u2026\"   57 minutes ago   Up 36 minutes (healthy)                                                                                     ceph-ansible",
                                "c3b8f043a57a   registry.osism.tech/osism/ara-server:1.7.3                      \"sh -c '/wait && /ru\u2026\"   57 minutes ago   Up 36 minutes (healthy)   8000/tcp                                                                          manager-ara-server-1",
                                "36eb7f15777c   registry.osism.tech/osism/osism-frontend:latest                 \"docker-entrypoint.s\u2026\"   57 minutes ago   Up 37 minutes             192.168.16.5:3000->3000/tcp                                                       osism-frontend",
                                "2cfcad7bb404   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   57 minutes ago   Up 37 minutes (healthy)                                                                                     manager-openstack-1",
                                "538cf49a52a2   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   57 minutes ago   Up 37 minutes (healthy)                                                                                     manager-listener-1",
                                "8cc90e9c6172   registry.osism.tech/dockerhub/library/redis:7.4.7-alpine        \"docker-entrypoint.s\u2026\"   57 minutes ago   Up 37 minutes (healthy)   6379/tcp                                                                          manager-redis-1",
                                "72d89da16cba   registry.osism.tech/dockerhub/library/mariadb:11.8.4            \"docker-entrypoint.s\u2026\"   57 minutes ago   Up 37 minutes (healthy)   3306/tcp                                                                          manager-mariadb-1",
                                "2d9151476f4b   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   57 minutes ago   Up 37 minutes (healthy)                                                                                     manager-beat-1",
                                "2f67b711771f   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- sleep\u2026\"   57 minutes ago   Up 37 minutes (healthy)                                                                                     osismclient",
                                "c309a338448f   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   57 minutes ago   Up 37 minutes (healthy)                                                                                     manager-flower-1",
                                "bc21652d95e6   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   57 minutes ago   Up 37 minutes (healthy)   192.168.16.5:8000->8000/tcp                                                       manager-api-1",
                                "12dae91709a5   registry.osism.tech/dockerhub/library/traefik:v3.5.0            \"/entrypoint.sh trae\u2026\"   58 minutes ago   Up 58 minutes (healthy)   192.168.16.5:80->80/tcp, 192.168.16.5:443->443/tcp, 192.168.16.5:8122->8080/tcp   traefik",
                                "",
                                "## Images @ testbed-manager",
                                "",
                                "+ echo",
                                "+ echo '## Images @ testbed-manager'",
                                "+ echo",
                                "+ osism container testbed-manager images",
                                "REPOSITORY                                               TAG              IMAGE ID       CREATED             SIZE",
                                "registry.osism.tech/osism/osism-ansible                  latest           5c181d8e29f7   About an hour ago   634MB",
                                "registry.osism.tech/osism/osism-kubernetes               latest           eb3836cac167   About an hour ago   1.24GB",
                                "registry.osism.tech/osism/kolla-ansible                  2024.2           6a7c8613a511   About an hour ago   631MB",
                                "registry.osism.tech/osism/ceph-ansible                   reef             ff5b1d600058   About an hour ago   581MB",
                                "registry.osism.tech/osism/osism                          latest           4102582d9e9d   About an hour ago   408MB",
                                "registry.osism.tech/osism/osism-frontend                 latest           fd751f4a01e0   About an hour ago   232MB",
                                "registry.osism.tech/osism/inventory-reconciler           latest           056df03054cc   About an hour ago   357MB",
                                "registry.osism.tech/osism/openstackclient                2024.2           87a2e78155c3   21 hours ago        239MB",
                                "registry.osism.tech/osism/cephclient                     reef             102963fef6c9   21 hours ago        453MB",
                                "registry.osism.tech/kolla/kolla-toolbox                  2024.2           2ff405629e62   23 hours ago        677MB",
                                "registry.osism.tech/kolla/cron                           2024.2           23ec8544c793   23 hours ago        275MB",
                                "registry.osism.tech/kolla/fluentd                        2024.2           851a383a0141   23 hours ago        589MB",
                                "registry.osism.tech/kolla/prometheus-alertmanager        2024.2           eb2991a15b6c   24 hours ago        413MB",
                                "registry.osism.tech/kolla/prometheus-node-exporter       2024.2           e9fbd7018fbf   24 hours ago        315MB",
                                "registry.osism.tech/kolla/prometheus-cadvisor            2024.2           1520e18eb877   24 hours ago        367MB",
                                "registry.osism.tech/kolla/prometheus-blackbox-exporter   2024.2           f23bcd4578b3   24 hours ago        317MB",
                                "registry.osism.tech/kolla/prometheus-v2-server           2024.2           ecfb9f783b6e   24 hours ago        849MB",
                                "registry.osism.tech/dockerhub/library/redis              7.4.7-alpine     e08bd8d5a677   7 weeks ago         41.4MB",
                                "registry.osism.tech/osism/homer                          v25.10.1         ea34b371c716   3 months ago        11.5MB",
                                "registry.osism.tech/dockerhub/library/mariadb            11.8.4           70745dd8f1d0   4 months ago        334MB",
                                "phpmyadmin/phpmyadmin                                    5.2              e66b1f5a8c58   5 months ago        742MB",
                                "registry.osism.tech/osism/ara-server                     1.7.3            d1b687333f2f   6 months ago        275MB",
                                "registry.osism.tech/dockerhub/library/traefik            v3.5.0           11cc59587f6a   8 months ago        226MB",
                                "registry.osism.tech/dockerhub/ubuntu/squid               6.1-23.10_beta   34b6bbbcf74b   21 months ago       146MB",
                                "+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2",
                                "++ semver latest 5.0.0",
                                "",
                                "## Containers @ testbed-node-0",
                                "",
                                "+ [[ -1 -eq -1 ]]",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ echo",
                                "+ echo '## Containers @ testbed-node-0'",
                                "+ echo",
                                "+ osism container testbed-node-0 ps",
                                "CONTAINER ID   IMAGE                                                                COMMAND                  CREATED          STATUS                    PORTS     NAMES",
                                "9d8922e858b5   registry.osism.tech/kolla/octavia-worker:2024.2                      \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_worker",
                                "a6ee23841317   registry.osism.tech/kolla/octavia-housekeeping:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_housekeeping",
                                "4a5db920cf4d   registry.osism.tech/kolla/octavia-health-manager:2024.2              \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_health_manager",
                                "8afd7c71e10e   registry.osism.tech/kolla/octavia-driver-agent:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes                        octavia_driver_agent",
                                "d75168b985b3   registry.osism.tech/kolla/octavia-api:2024.2                         \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_api",
                                "112fca436d51   registry.osism.tech/kolla/nova-novncproxy:2024.2                     \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_novncproxy",
                                "2debb8ac5833   registry.osism.tech/kolla/nova-conductor:2024.2                      \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_conductor",
                                "17b1c2bc2a93   registry.osism.tech/kolla/nova-api:2024.2                            \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_api",
                                "d03f58f2a91a   registry.osism.tech/kolla/nova-scheduler:2024.2                      \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_scheduler",
                                "079169f0167c   registry.osism.tech/kolla/grafana:2024.2                             \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes                        grafana",
                                "9563220c0f32   registry.osism.tech/kolla/cinder-backup:2024.2                       \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_backup",
                                "01f462bda636   registry.osism.tech/kolla/cinder-volume:2024.2                       \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_volume",
                                "7167a3f3c3b3   registry.osism.tech/kolla/glance-api:2024.2                          \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             glance_api",
                                "670c1d20efa4   registry.osism.tech/kolla/cinder-scheduler:2024.2                    \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_scheduler",
                                "789c5c06779c   registry.osism.tech/kolla/cinder-api:2024.2                          \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_api",
                                "819bdc497753   registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2   \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_elasticsearch_exporter",
                                "ba617ee789c0   registry.osism.tech/kolla/prometheus-cadvisor:2024.2                 \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_cadvisor",
                                "9eab234cd872   registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_memcached_exporter",
                                "1c50f7207750   registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2          \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_mysqld_exporter",
                                "d081c69342ac   registry.osism.tech/kolla/prometheus-node-exporter:2024.2            \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_node_exporter",
                                "60ef4c825ac7   registry.osism.tech/kolla/magnum-conductor:2024.2                    \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             magnum_conductor",
                                "0b1d6881c3b6   registry.osism.tech/kolla/magnum-api:2024.2                          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             magnum_api",
                                "2a8d8d114469   registry.osism.tech/kolla/neutron-server:2024.2                      \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             neutron_server",
                                "0889cecb0ed2   registry.osism.tech/kolla/designate-worker:2024.2                    \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             designate_worker",
                                "23631501f2a2   registry.osism.tech/kolla/placement-api:2024.2                       \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             placement_api",
                                "85a3b0d8dd80   registry.osism.tech/kolla/designate-mdns:2024.2                      \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_mdns",
                                "2facf6912789   registry.osism.tech/kolla/designate-producer:2024.2                  \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_producer",
                                "9690c76363a7   registry.osism.tech/kolla/designate-central:2024.2                   \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_central",
                                "6440e8936ce4   registry.osism.tech/kolla/designate-api:2024.2                       \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_api",
                                "5d02b1d19aec   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mgr -\u2026\"   14 minutes ago   Up 14 minutes                       ceph-mgr-testbed-node-0",
                                "8ccd851e0dfe   registry.osism.tech/kolla/barbican-worker:2024.2                     \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             barbican_worker",
                                "111d3ae52613   registry.osism.tech/kolla/designate-backend-bind9:2024.2             \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_backend_bind9",
                                "102484264d0e   registry.osism.tech/kolla/barbican-keystone-listener:2024.2          \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             barbican_keystone_listener",
                                "e10dd320f39d   registry.osism.tech/kolla/barbican-api:2024.2                        \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             barbican_api",
                                "3425e4fe5ea9   registry.osism.tech/kolla/keystone:2024.2                            \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             keystone",
                                "19fdb3f3eae3   registry.osism.tech/kolla/keystone-fernet:2024.2                     \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             keystone_fernet",
                                "abaa7f710593   registry.osism.tech/kolla/keystone-ssh:2024.2                        \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_ssh",
                                "e75cf675d062   registry.osism.tech/kolla/horizon:2024.2                             \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             horizon",
                                "70eba282b12c   registry.osism.tech/kolla/mariadb-server:2024.2                      \"dumb-init -- kolla_\u2026\"   19 minutes ago   Up 19 minutes (healthy)             mariadb",
                                "38749466ddbb   registry.osism.tech/kolla/opensearch-dashboards:2024.2               \"dumb-init --single-\u2026\"   20 minutes ago   Up 20 minutes (healthy)             opensearch_dashboards",
                                "4b878a379f89   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-crash\"    21 minutes ago   Up 21 minutes                       ceph-crash-testbed-node-0",
                                "06727c8ef341   registry.osism.tech/kolla/opensearch:2024.2                          \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             opensearch",
                                "e6cd19bfe23a   registry.osism.tech/kolla/keepalived:2024.2                          \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes                       keepalived",
                                "bd8a3271595a   registry.osism.tech/kolla/proxysql:2024.2                            \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes (healthy)             proxysql",
                                "d3d6d8a0896b   registry.osism.tech/kolla/haproxy:2024.2                             \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes (healthy)             haproxy",
                                "60aef232a911   registry.osism.tech/kolla/ovn-northd:2024.2                          \"dumb-init --single-\u2026\"   25 minutes ago   Up 25 minutes                       ovn_northd",
                                "c5e129f45a17   registry.osism.tech/kolla/ovn-sb-db-server:2024.2                    \"dumb-init --single-\u2026\"   25 minutes ago   Up 25 minutes                       ovn_sb_db",
                                "ebdab83132e9   registry.osism.tech/kolla/ovn-nb-db-server:2024.2                    \"dumb-init --single-\u2026\"   25 minutes ago   Up 25 minutes                       ovn_nb_db",
                                "e75f5e68cdc6   registry.osism.tech/kolla/ovn-controller:2024.2                      \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       ovn_controller",
                                "22929cdeac3a   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mon -\u2026\"   26 minutes ago   Up 26 minutes                       ceph-mon-testbed-node-0",
                                "9f9c1ac5d698   registry.osism.tech/kolla/rabbitmq:2024.2                            \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             rabbitmq",
                                "50d0a4f1eb1d   registry.osism.tech/kolla/openvswitch-vswitchd:2024.2                \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             openvswitch_vswitchd",
                                "c27b35eea4c1   registry.osism.tech/kolla/openvswitch-db-server:2024.2               \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             openvswitch_db",
                                "d7ac3868d9d8   registry.osism.tech/kolla/redis-sentinel:2024.2                      \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             redis_sentinel",
                                "30b4cc84749b   registry.osism.tech/kolla/redis:2024.2                               \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             redis",
                                "759d15301192   registry.osism.tech/kolla/memcached:2024.2                           \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             memcached",
                                "1ccc23d1ce4f   registry.osism.tech/kolla/cron:2024.2                                \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes                       cron",
                                "d32c9516f38b   registry.osism.tech/kolla/kolla-toolbox:2024.2                       \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                       kolla_toolbox",
                                "de86538f2392   registry.osism.tech/kolla/fluentd:2024.2                             \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                       fluentd",
                                "",
                                "## Images @ testbed-node-0",
                                "",
                                "+ echo",
                                "+ echo '## Images @ testbed-node-0'",
                                "+ echo",
                                "+ osism container testbed-node-0 images",
                                "REPOSITORY                                                    TAG       IMAGE ID       CREATED        SIZE",
                                "registry.osism.tech/osism/ceph-daemon                         reef      60d930974fb1   22 hours ago   1.27GB",
                                "registry.osism.tech/kolla/keepalived                          2024.2    b2b59a47e345   23 hours ago   286MB",
                                "registry.osism.tech/kolla/grafana                             2024.2    25f6c06d0021   23 hours ago   1.04GB",
                                "registry.osism.tech/kolla/kolla-toolbox                       2024.2    2ff405629e62   23 hours ago   677MB",
                                "registry.osism.tech/kolla/cron                                2024.2    23ec8544c793   23 hours ago   275MB",
                                "registry.osism.tech/kolla/opensearch-dashboards               2024.2    687dcde2bcb8   23 hours ago   1.54GB",
                                "registry.osism.tech/kolla/opensearch                          2024.2    b409ad33e488   23 hours ago   1.57GB",
                                "registry.osism.tech/kolla/memcached                           2024.2    ed2bff2a6973   23 hours ago   276MB",
                                "registry.osism.tech/kolla/proxysql                            2024.2    b4d6db8fd341   23 hours ago   426MB",
                                "registry.osism.tech/kolla/fluentd                             2024.2    851a383a0141   23 hours ago   589MB",
                                "registry.osism.tech/kolla/rabbitmq                            2024.2    a4ab33fd70fd   23 hours ago   332MB",
                                "registry.osism.tech/kolla/haproxy                             2024.2    3fa2a10107a8   23 hours ago   284MB",
                                "registry.osism.tech/kolla/redis-sentinel                      2024.2    5f4f04f4383c   24 hours ago   282MB",
                                "registry.osism.tech/kolla/redis                               2024.2    533c487568ee   24 hours ago   282MB",
                                "registry.osism.tech/kolla/mariadb-server                      2024.2    297f8bb339a1   24 hours ago   462MB",
                                "registry.osism.tech/kolla/prometheus-memcached-exporter       2024.2    dd336c43b779   24 hours ago   308MB",
                                "registry.osism.tech/kolla/prometheus-node-exporter            2024.2    e9fbd7018fbf   24 hours ago   315MB",
                                "registry.osism.tech/kolla/prometheus-mysqld-exporter          2024.2    548293489d76   24 hours ago   310MB",
                                "registry.osism.tech/kolla/prometheus-cadvisor                 2024.2    1520e18eb877   24 hours ago   367MB",
                                "registry.osism.tech/kolla/prometheus-elasticsearch-exporter   2024.2    5b796764074a   24 hours ago   301MB",
                                "registry.osism.tech/kolla/horizon                             2024.2    50eeb61edd36   24 hours ago   1.16GB",
                                "registry.osism.tech/kolla/openvswitch-vswitchd                2024.2    dbcb0c63d3a2   24 hours ago   288MB",
                                "registry.osism.tech/kolla/openvswitch-db-server               2024.2    a8f2511eff1f   24 hours ago   288MB",
                                "registry.osism.tech/kolla/neutron-server                      2024.2    7c6811afdb1b   24 hours ago   1.17GB",
                                "registry.osism.tech/kolla/octavia-api                         2024.2    eaf97bcdf081   24 hours ago   1.06GB",
                                "registry.osism.tech/kolla/octavia-driver-agent                2024.2    4c4abbcba868   24 hours ago   1.06GB",
                                "registry.osism.tech/kolla/octavia-worker                      2024.2    ae8999de916c   24 hours ago   1.04GB",
                                "registry.osism.tech/kolla/octavia-housekeeping                2024.2    e15b03944567   24 hours ago   1.04GB",
                                "registry.osism.tech/kolla/octavia-health-manager              2024.2    41dcecaeaf89   24 hours ago   1.04GB",
                                "registry.osism.tech/kolla/skyline-apiserver                   2024.2    9f9c69d19c31   24 hours ago   999MB",
                                "registry.osism.tech/kolla/skyline-console                     2024.2    f5b7b069c854   24 hours ago   1.06GB",
                                "registry.osism.tech/kolla/keystone                            2024.2    a6b20f41325d   24 hours ago   1.08GB",
                                "registry.osism.tech/kolla/keystone-fernet                     2024.2    15aaf5c48263   24 hours ago   1.05GB",
                                "registry.osism.tech/kolla/keystone-ssh                        2024.2    2634cde67cfd   24 hours ago   1.05GB",
                                "registry.osism.tech/kolla/cinder-backup                       2024.2    5f8e22fdae27   24 hours ago   1.42GB",
                                "registry.osism.tech/kolla/cinder-volume                       2024.2    29801e980d03   24 hours ago   1.73GB",
                                "registry.osism.tech/kolla/cinder-scheduler                    2024.2    7d98f6182a5d   24 hours ago   1.41GB",
                                "registry.osism.tech/kolla/cinder-api                          2024.2    c16461b7e25c   24 hours ago   1.41GB",
                                "registry.osism.tech/kolla/barbican-worker                     2024.2    cbe9a09e642e   24 hours ago   1GB",
                                "registry.osism.tech/kolla/barbican-keystone-listener          2024.2    8b79c739c5aa   24 hours ago   1GB",
                                "registry.osism.tech/kolla/barbican-api                        2024.2    ec75be03f45f   24 hours ago   1e+03MB",
                                "registry.osism.tech/kolla/nova-scheduler                      2024.2    d13b758badcb   24 hours ago   1.22GB",
                                "registry.osism.tech/kolla/nova-api                            2024.2    fa74d9e1d80c   24 hours ago   1.22GB",
                                "registry.osism.tech/kolla/nova-novncproxy                     2024.2    b0f517fc2110   24 hours ago   1.38GB",
                                "registry.osism.tech/kolla/nova-conductor                      2024.2    cc5099f460a9   24 hours ago   1.22GB",
                                "registry.osism.tech/kolla/ceilometer-notification             2024.2    447d352c2733   24 hours ago   985MB",
                                "registry.osism.tech/kolla/ceilometer-central                  2024.2    494e092cff9e   24 hours ago   986MB",
                                "registry.osism.tech/kolla/glance-api                          2024.2    0c1d9d7160de   24 hours ago   1.11GB",
                                "registry.osism.tech/kolla/magnum-conductor                    2024.2    e4d602a3ad04   24 hours ago   1.25GB",
                                "registry.osism.tech/kolla/magnum-api                          2024.2    083abdbaa2fa   24 hours ago   1.14GB",
                                "registry.osism.tech/kolla/aodh-evaluator                      2024.2    e38bf12d2f87   24 hours ago   984MB",
                                "registry.osism.tech/kolla/aodh-notifier                       2024.2    d90172096345   24 hours ago   984MB",
                                "registry.osism.tech/kolla/aodh-listener                       2024.2    9cd4a42b1c66   24 hours ago   984MB",
                                "registry.osism.tech/kolla/aodh-api                            2024.2    ca349612b64e   24 hours ago   983MB",
                                "registry.osism.tech/kolla/placement-api                       2024.2    dfc171655fae   24 hours ago   985MB",
                                "registry.osism.tech/kolla/designate-worker                    2024.2    e9363cf6baa5   24 hours ago   998MB",
                                "registry.osism.tech/kolla/designate-producer                  2024.2    1d3f00ae0cb0   24 hours ago   994MB",
                                "registry.osism.tech/kolla/designate-mdns                      2024.2    b74dea85ff31   24 hours ago   994MB",
                                "registry.osism.tech/kolla/designate-backend-bind9             2024.2    71020f79d0b8   24 hours ago   998MB",
                                "registry.osism.tech/kolla/designate-central                   2024.2    e270fdf55dd9   24 hours ago   993MB",
                                "registry.osism.tech/kolla/designate-api                       2024.2    2b71cfe02281   24 hours ago   994MB",
                                "registry.osism.tech/kolla/ovn-nb-db-server                    2024.2    91492a927f10   24 hours ago   850MB",
                                "registry.osism.tech/kolla/ovn-sb-db-server                    2024.2    c3d787fcf696   24 hours ago   850MB",
                                "registry.osism.tech/kolla/ovn-controller                      2024.2    df4830242254   24 hours ago   850MB",
                                "registry.osism.tech/kolla/ovn-northd                          2024.2    521cb38ac60e   24 hours ago   850MB",
                                "+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2",
                                "++ semver latest 5.0.0",
                                "",
                                "## Containers @ testbed-node-1",
                                "",
                                "+ [[ -1 -eq -1 ]]",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ echo",
                                "+ echo '## Containers @ testbed-node-1'",
                                "+ echo",
                                "+ osism container testbed-node-1 ps",
                                "CONTAINER ID   IMAGE                                                                COMMAND                  CREATED          STATUS                    PORTS     NAMES",
                                "b96d6747a23c   registry.osism.tech/kolla/octavia-worker:2024.2                      \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_worker",
                                "e13afae5aa0a   registry.osism.tech/kolla/octavia-housekeeping:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_housekeeping",
                                "da03c3f203d8   registry.osism.tech/kolla/octavia-health-manager:2024.2              \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_health_manager",
                                "9cfb28dcd39f   registry.osism.tech/kolla/octavia-driver-agent:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes                        octavia_driver_agent",
                                "62a640b1abe4   registry.osism.tech/kolla/octavia-api:2024.2                         \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_api",
                                "42fbd84ca280   registry.osism.tech/kolla/nova-novncproxy:2024.2                     \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_novncproxy",
                                "b5219b0892fb   registry.osism.tech/kolla/nova-conductor:2024.2                      \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_conductor",
                                "d0994dd06228   registry.osism.tech/kolla/grafana:2024.2                             \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes                        grafana",
                                "fa5a07dbd841   registry.osism.tech/kolla/nova-api:2024.2                            \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_api",
                                "708a6fe44ce6   registry.osism.tech/kolla/nova-scheduler:2024.2                      \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_scheduler",
                                "fc9c5048ab2c   registry.osism.tech/kolla/cinder-backup:2024.2                       \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_backup",
                                "6c3ee47cae97   registry.osism.tech/kolla/cinder-volume:2024.2                       \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_volume",
                                "078677f3587a   registry.osism.tech/kolla/glance-api:2024.2                          \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             glance_api",
                                "b9fae0b403e3   registry.osism.tech/kolla/cinder-scheduler:2024.2                    \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_scheduler",
                                "5683d8da194d   registry.osism.tech/kolla/cinder-api:2024.2                          \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_api",
                                "c671e3dd9cbb   registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2   \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_elasticsearch_exporter",
                                "43fd6bbce2aa   registry.osism.tech/kolla/prometheus-cadvisor:2024.2                 \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_cadvisor",
                                "d9fc738086bb   registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_memcached_exporter",
                                "9761154a7e32   registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2          \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_mysqld_exporter",
                                "54125ce3e08c   registry.osism.tech/kolla/prometheus-node-exporter:2024.2            \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_node_exporter",
                                "d4a0b0e8dac8   registry.osism.tech/kolla/magnum-conductor:2024.2                    \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             magnum_conductor",
                                "3e1bb59dea40   registry.osism.tech/kolla/magnum-api:2024.2                          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             magnum_api",
                                "8ad6908d91f0   registry.osism.tech/kolla/neutron-server:2024.2                      \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             neutron_server",
                                "12d4f9f919c3   registry.osism.tech/kolla/designate-worker:2024.2                    \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             designate_worker",
                                "21f63530e59c   registry.osism.tech/kolla/placement-api:2024.2                       \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             placement_api",
                                "40eeee9a200e   registry.osism.tech/kolla/designate-mdns:2024.2                      \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_mdns",
                                "086862740b99   registry.osism.tech/kolla/designate-producer:2024.2                  \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_producer",
                                "4c7bdbda027f   registry.osism.tech/kolla/designate-central:2024.2                   \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_central",
                                "79982be92a5e   registry.osism.tech/kolla/designate-api:2024.2                       \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_api",
                                "58f711bf95fb   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mgr -\u2026\"   14 minutes ago   Up 14 minutes                       ceph-mgr-testbed-node-1",
                                "0aa25d6997ac   registry.osism.tech/kolla/barbican-worker:2024.2                     \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             barbican_worker",
                                "e0df55fbd5d2   registry.osism.tech/kolla/designate-backend-bind9:2024.2             \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_backend_bind9",
                                "e4902252a4d7   registry.osism.tech/kolla/barbican-keystone-listener:2024.2          \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             barbican_keystone_listener",
                                "2ee84b1a3de7   registry.osism.tech/kolla/barbican-api:2024.2                        \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             barbican_api",
                                "fad8b7e4e4f8   registry.osism.tech/kolla/keystone:2024.2                            \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             keystone",
                                "be98159212c6   registry.osism.tech/kolla/keystone-fernet:2024.2                     \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             keystone_fernet",
                                "8391e6c5b1b4   registry.osism.tech/kolla/horizon:2024.2                             \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             horizon",
                                "026edb57c8f2   registry.osism.tech/kolla/keystone-ssh:2024.2                        \"dumb-init --single-\u2026\"   18 minutes ago   Up 17 minutes (healthy)             keystone_ssh",
                                "e3d0530f6f13   registry.osism.tech/kolla/opensearch-dashboards:2024.2               \"dumb-init --single-\u2026\"   19 minutes ago   Up 19 minutes (healthy)             opensearch_dashboards",
                                "903dca551e29   registry.osism.tech/kolla/mariadb-server:2024.2                      \"dumb-init -- kolla_\u2026\"   20 minutes ago   Up 20 minutes (healthy)             mariadb",
                                "abef70983a85   registry.osism.tech/kolla/opensearch:2024.2                          \"dumb-init --single-\u2026\"   20 minutes ago   Up 20 minutes (healthy)             opensearch",
                                "4caa03a44222   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-crash\"    21 minutes ago   Up 21 minutes                       ceph-crash-testbed-node-1",
                                "bd27d8a16f20   registry.osism.tech/kolla/keepalived:2024.2                          \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes                       keepalived",
                                "76f86a34f0e0   registry.osism.tech/kolla/proxysql:2024.2                            \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes (healthy)             proxysql",
                                "6deaa4399b35   registry.osism.tech/kolla/haproxy:2024.2                             \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes (healthy)             haproxy",
                                "2e8697946763   registry.osism.tech/kolla/ovn-northd:2024.2                          \"dumb-init --single-\u2026\"   25 minutes ago   Up 24 minutes                       ovn_northd",
                                "afc85eabf323   registry.osism.tech/kolla/ovn-sb-db-server:2024.2                    \"dumb-init --single-\u2026\"   25 minutes ago   Up 24 minutes                       ovn_sb_db",
                                "5c29f22f1712   registry.osism.tech/kolla/ovn-nb-db-server:2024.2                    \"dumb-init --single-\u2026\"   25 minutes ago   Up 24 minutes                       ovn_nb_db",
                                "2c89399c74be   registry.osism.tech/kolla/ovn-controller:2024.2                      \"dumb-init --single-\u2026\"   25 minutes ago   Up 25 minutes                       ovn_controller",
                                "ee2851244d11   registry.osism.tech/kolla/rabbitmq:2024.2                            \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes (healthy)             rabbitmq",
                                "a96b2c5299e5   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mon -\u2026\"   26 minutes ago   Up 26 minutes                       ceph-mon-testbed-node-1",
                                "f7f9ce0f9b1c   registry.osism.tech/kolla/openvswitch-vswitchd:2024.2                \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             openvswitch_vswitchd",
                                "3f280ab09ce9   registry.osism.tech/kolla/openvswitch-db-server:2024.2               \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             openvswitch_db",
                                "1077e37f68a8   registry.osism.tech/kolla/redis-sentinel:2024.2                      \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             redis_sentinel",
                                "8fb396a44b47   registry.osism.tech/kolla/redis:2024.2                               \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             redis",
                                "2ed0203510f4   registry.osism.tech/kolla/memcached:2024.2                           \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             memcached",
                                "7631a1302929   registry.osism.tech/kolla/cron:2024.2                                \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes                       cron",
                                "0399bb27a879   registry.osism.tech/kolla/kolla-toolbox:2024.2                       \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes                       kolla_toolbox",
                                "6618cc36166a   registry.osism.tech/kolla/fluentd:2024.2                             \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                       fluentd",
                                "",
                                "## Images @ testbed-node-1",
                                "",
                                "+ echo",
                                "+ echo '## Images @ testbed-node-1'",
                                "+ echo",
                                "+ osism container testbed-node-1 images",
                                "REPOSITORY                                                    TAG       IMAGE ID       CREATED        SIZE",
                                "registry.osism.tech/osism/ceph-daemon                         reef      60d930974fb1   22 hours ago   1.27GB",
                                "registry.osism.tech/kolla/keepalived                          2024.2    b2b59a47e345   23 hours ago   286MB",
                                "registry.osism.tech/kolla/grafana                             2024.2    25f6c06d0021   23 hours ago   1.04GB",
                                "registry.osism.tech/kolla/kolla-toolbox                       2024.2    2ff405629e62   23 hours ago   677MB",
                                "registry.osism.tech/kolla/cron                                2024.2    23ec8544c793   23 hours ago   275MB",
                                "registry.osism.tech/kolla/opensearch-dashboards               2024.2    687dcde2bcb8   23 hours ago   1.54GB",
                                "registry.osism.tech/kolla/opensearch                          2024.2    b409ad33e488   23 hours ago   1.57GB",
                                "registry.osism.tech/kolla/memcached                           2024.2    ed2bff2a6973   23 hours ago   276MB",
                                "registry.osism.tech/kolla/proxysql                            2024.2    b4d6db8fd341   23 hours ago   426MB",
                                "registry.osism.tech/kolla/fluentd                             2024.2    851a383a0141   23 hours ago   589MB",
                                "registry.osism.tech/kolla/rabbitmq                            2024.2    a4ab33fd70fd   23 hours ago   332MB",
                                "registry.osism.tech/kolla/haproxy                             2024.2    3fa2a10107a8   24 hours ago   284MB",
                                "registry.osism.tech/kolla/redis-sentinel                      2024.2    5f4f04f4383c   24 hours ago   282MB",
                                "registry.osism.tech/kolla/redis                               2024.2    533c487568ee   24 hours ago   282MB",
                                "registry.osism.tech/kolla/mariadb-server                      2024.2    297f8bb339a1   24 hours ago   462MB",
                                "registry.osism.tech/kolla/prometheus-memcached-exporter       2024.2    dd336c43b779   24 hours ago   308MB",
                                "registry.osism.tech/kolla/prometheus-node-exporter            2024.2    e9fbd7018fbf   24 hours ago   315MB",
                                "registry.osism.tech/kolla/prometheus-mysqld-exporter          2024.2    548293489d76   24 hours ago   310MB",
                                "registry.osism.tech/kolla/prometheus-cadvisor                 2024.2    1520e18eb877   24 hours ago   367MB",
                                "registry.osism.tech/kolla/prometheus-elasticsearch-exporter   2024.2    5b796764074a   24 hours ago   301MB",
                                "registry.osism.tech/kolla/horizon                             2024.2    50eeb61edd36   24 hours ago   1.16GB",
                                "registry.osism.tech/kolla/openvswitch-vswitchd                2024.2    dbcb0c63d3a2   24 hours ago   288MB",
                                "registry.osism.tech/kolla/openvswitch-db-server               2024.2    a8f2511eff1f   24 hours ago   288MB",
                                "registry.osism.tech/kolla/neutron-server                      2024.2    7c6811afdb1b   24 hours ago   1.17GB",
                                "registry.osism.tech/kolla/octavia-api                         2024.2    eaf97bcdf081   24 hours ago   1.06GB",
                                "registry.osism.tech/kolla/octavia-driver-agent                2024.2    4c4abbcba868   24 hours ago   1.06GB",
                                "registry.osism.tech/kolla/octavia-worker                      2024.2    ae8999de916c   24 hours ago   1.04GB",
                                "registry.osism.tech/kolla/octavia-housekeeping                2024.2    e15b03944567   24 hours ago   1.04GB",
                                "registry.osism.tech/kolla/octavia-health-manager              2024.2    41dcecaeaf89   24 hours ago   1.04GB",
                                "registry.osism.tech/kolla/keystone                            2024.2    a6b20f41325d   24 hours ago   1.08GB",
                                "registry.osism.tech/kolla/keystone-fernet                     2024.2    15aaf5c48263   24 hours ago   1.05GB",
                                "registry.osism.tech/kolla/keystone-ssh                        2024.2    2634cde67cfd   24 hours ago   1.05GB",
                                "registry.osism.tech/kolla/cinder-backup                       2024.2    5f8e22fdae27   24 hours ago   1.42GB",
                                "registry.osism.tech/kolla/cinder-volume                       2024.2    29801e980d03   24 hours ago   1.73GB",
                                "registry.osism.tech/kolla/cinder-scheduler                    2024.2    7d98f6182a5d   24 hours ago   1.41GB",
                                "registry.osism.tech/kolla/cinder-api                          2024.2    c16461b7e25c   24 hours ago   1.41GB",
                                "registry.osism.tech/kolla/barbican-worker                     2024.2    cbe9a09e642e   24 hours ago   1GB",
                                "registry.osism.tech/kolla/barbican-keystone-listener          2024.2    8b79c739c5aa   24 hours ago   1GB",
                                "registry.osism.tech/kolla/barbican-api                        2024.2    ec75be03f45f   24 hours ago   1e+03MB",
                                "registry.osism.tech/kolla/nova-scheduler                      2024.2    d13b758badcb   24 hours ago   1.22GB",
                                "registry.osism.tech/kolla/nova-api                            2024.2    fa74d9e1d80c   24 hours ago   1.22GB",
                                "registry.osism.tech/kolla/nova-novncproxy                     2024.2    b0f517fc2110   24 hours ago   1.38GB",
                                "registry.osism.tech/kolla/nova-conductor                      2024.2    cc5099f460a9   24 hours ago   1.22GB",
                                "registry.osism.tech/kolla/glance-api                          2024.2    0c1d9d7160de   24 hours ago   1.11GB",
                                "registry.osism.tech/kolla/magnum-conductor                    2024.2    e4d602a3ad04   24 hours ago   1.25GB",
                                "registry.osism.tech/kolla/magnum-api                          2024.2    083abdbaa2fa   24 hours ago   1.14GB",
                                "registry.osism.tech/kolla/placement-api                       2024.2    dfc171655fae   24 hours ago   985MB",
                                "registry.osism.tech/kolla/designate-worker                    2024.2    e9363cf6baa5   24 hours ago   998MB",
                                "registry.osism.tech/kolla/designate-producer                  2024.2    1d3f00ae0cb0   24 hours ago   994MB",
                                "registry.osism.tech/kolla/designate-mdns                      2024.2    b74dea85ff31   24 hours ago   994MB",
                                "registry.osism.tech/kolla/designate-backend-bind9             2024.2    71020f79d0b8   24 hours ago   998MB",
                                "registry.osism.tech/kolla/designate-central                   2024.2    e270fdf55dd9   24 hours ago   993MB",
                                "registry.osism.tech/kolla/designate-api                       2024.2    2b71cfe02281   24 hours ago   994MB",
                                "registry.osism.tech/kolla/ovn-sb-db-server                    2024.2    c3d787fcf696   24 hours ago   850MB",
                                "registry.osism.tech/kolla/ovn-nb-db-server                    2024.2    91492a927f10   24 hours ago   850MB",
                                "registry.osism.tech/kolla/ovn-controller                      2024.2    df4830242254   24 hours ago   850MB",
                                "registry.osism.tech/kolla/ovn-northd                          2024.2    521cb38ac60e   24 hours ago   850MB",
                                "+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2",
                                "++ semver latest 5.0.0",
                                "",
                                "## Containers @ testbed-node-2",
                                "",
                                "+ [[ -1 -eq -1 ]]",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ echo",
                                "+ echo '## Containers @ testbed-node-2'",
                                "+ echo",
                                "+ osism container testbed-node-2 ps",
                                "CONTAINER ID   IMAGE                                                                COMMAND                  CREATED          STATUS                    PORTS     NAMES",
                                "0ae5e2e46239   registry.osism.tech/kolla/octavia-worker:2024.2                      \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_worker",
                                "ac59fbdcb5c0   registry.osism.tech/kolla/octavia-housekeeping:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_housekeeping",
                                "38f548bbc74b   registry.osism.tech/kolla/octavia-health-manager:2024.2              \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_health_manager",
                                "0dcc5a926d5b   registry.osism.tech/kolla/octavia-driver-agent:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes                        octavia_driver_agent",
                                "8a02191defe9   registry.osism.tech/kolla/octavia-api:2024.2                         \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_api",
                                "39264437dc74   registry.osism.tech/kolla/nova-novncproxy:2024.2                     \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_novncproxy",
                                "f42b2ff11f0b   registry.osism.tech/kolla/nova-conductor:2024.2                      \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_conductor",
                                "18b5f3a218c9   registry.osism.tech/kolla/grafana:2024.2                             \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes                        grafana",
                                "125c018b939c   registry.osism.tech/kolla/nova-api:2024.2                            \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_api",
                                "40c1053d59c7   registry.osism.tech/kolla/nova-scheduler:2024.2                      \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_scheduler",
                                "f7af2ca7b532   registry.osism.tech/kolla/cinder-backup:2024.2                       \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_backup",
                                "69c2a2a2d036   registry.osism.tech/kolla/cinder-volume:2024.2                       \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_volume",
                                "a76993a8d64a   registry.osism.tech/kolla/glance-api:2024.2                          \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             glance_api",
                                "56495cd3659a   registry.osism.tech/kolla/cinder-scheduler:2024.2                    \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_scheduler",
                                "2f36c0d068ef   registry.osism.tech/kolla/cinder-api:2024.2                          \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_api",
                                "704a31a41152   registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2   \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_elasticsearch_exporter",
                                "42d0729cc93d   registry.osism.tech/kolla/prometheus-cadvisor:2024.2                 \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_cadvisor",
                                "88896184888a   registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_memcached_exporter",
                                "56ab80b6db9f   registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2          \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_mysqld_exporter",
                                "32fb2e56f1e2   registry.osism.tech/kolla/prometheus-node-exporter:2024.2            \"dumb-init --single-\u2026\"   12 minutes ago   Up 11 minutes                       prometheus_node_exporter",
                                "f1592e2057c5   registry.osism.tech/kolla/magnum-conductor:2024.2                    \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             magnum_conductor",
                                "722b223d64fb   registry.osism.tech/kolla/magnum-api:2024.2                          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             magnum_api",
                                "db28a4c0beec   registry.osism.tech/kolla/neutron-server:2024.2                      \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             neutron_server",
                                "967298c31eb9   registry.osism.tech/kolla/designate-worker:2024.2                    \"dumb-init --single-\u2026\"   14 minutes ago   Up 13 minutes (healthy)             designate_worker",
                                "1b4e69352242   registry.osism.tech/kolla/placement-api:2024.2                       \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             placement_api",
                                "61db02eb584f   registry.osism.tech/kolla/designate-mdns:2024.2                      \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_mdns",
                                "65134717cd42   registry.osism.tech/kolla/designate-producer:2024.2                  \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_producer",
                                "eb4673f5b25f   registry.osism.tech/kolla/designate-central:2024.2                   \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_central",
                                "0db26038f628   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mgr -\u2026\"   14 minutes ago   Up 14 minutes                       ceph-mgr-testbed-node-2",
                                "d1ae168b9a8d   registry.osism.tech/kolla/designate-api:2024.2                       \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_api",
                                "4a2814b312f1   registry.osism.tech/kolla/barbican-worker:2024.2                     \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             barbican_worker",
                                "dfd755dc6faa   registry.osism.tech/kolla/designate-backend-bind9:2024.2             \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_backend_bind9",
                                "0cb1920e7dd4   registry.osism.tech/kolla/barbican-keystone-listener:2024.2          \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             barbican_keystone_listener",
                                "3882321b1615   registry.osism.tech/kolla/barbican-api:2024.2                        \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_api",
                                "4ab3833b2116   registry.osism.tech/kolla/keystone:2024.2                            \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             keystone",
                                "f1f648fe902c   registry.osism.tech/kolla/keystone-fernet:2024.2                     \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             keystone_fernet",
                                "ef461c30c337   registry.osism.tech/kolla/horizon:2024.2                             \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             horizon",
                                "c234a31af7b6   registry.osism.tech/kolla/keystone-ssh:2024.2                        \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_ssh",
                                "2169ee164e49   registry.osism.tech/kolla/opensearch-dashboards:2024.2               \"dumb-init --single-\u2026\"   19 minutes ago   Up 19 minutes (healthy)             opensearch_dashboards",
                                "5a7def510321   registry.osism.tech/kolla/mariadb-server:2024.2                      \"dumb-init -- kolla_\u2026\"   20 minutes ago   Up 20 minutes (healthy)             mariadb",
                                "35628559acd2   registry.osism.tech/kolla/opensearch:2024.2                          \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             opensearch",
                                "a0ddbb326162   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-crash\"    21 minutes ago   Up 21 minutes                       ceph-crash-testbed-node-2",
                                "cb39873970ea   registry.osism.tech/kolla/keepalived:2024.2                          \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes                       keepalived",
                                "76a08b821526   registry.osism.tech/kolla/proxysql:2024.2                            \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes (healthy)             proxysql",
                                "e831ec38884e   registry.osism.tech/kolla/haproxy:2024.2                             \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes (healthy)             haproxy",
                                "4ba5cc59fc04   registry.osism.tech/kolla/ovn-northd:2024.2                          \"dumb-init --single-\u2026\"   25 minutes ago   Up 24 minutes                       ovn_northd",
                                "8e344c80c65a   registry.osism.tech/kolla/ovn-sb-db-server:2024.2                    \"dumb-init --single-\u2026\"   25 minutes ago   Up 24 minutes                       ovn_sb_db",
                                "18ffee7ce1bd   registry.osism.tech/kolla/ovn-nb-db-server:2024.2                    \"dumb-init --single-\u2026\"   25 minutes ago   Up 25 minutes                       ovn_nb_db",
                                "4ce96a10d15b   registry.osism.tech/kolla/rabbitmq:2024.2                            \"dumb-init --single-\u2026\"   25 minutes ago   Up 25 minutes (healthy)             rabbitmq",
                                "2194c60e76b2   registry.osism.tech/kolla/ovn-controller:2024.2                      \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       ovn_controller",
                                "57e0bd4375ae   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mon -\u2026\"   26 minutes ago   Up 26 minutes                       ceph-mon-testbed-node-2",
                                "99acbc14ede4   registry.osism.tech/kolla/openvswitch-vswitchd:2024.2                \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             openvswitch_vswitchd",
                                "9e5f2b62a875   registry.osism.tech/kolla/openvswitch-db-server:2024.2               \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             openvswitch_db",
                                "648e20862b4d   registry.osism.tech/kolla/redis-sentinel:2024.2                      \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             redis_sentinel",
                                "a32a1543446c   registry.osism.tech/kolla/redis:2024.2                               \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             redis",
                                "0b9bd0c5649d   registry.osism.tech/kolla/memcached:2024.2                           \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             memcached",
                                "5a710b5a3f25   registry.osism.tech/kolla/cron:2024.2                                \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes                       cron",
                                "b098600212de   registry.osism.tech/kolla/kolla-toolbox:2024.2                       \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes                       kolla_toolbox",
                                "c921c720e066   registry.osism.tech/kolla/fluentd:2024.2                             \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                       fluentd",
                                "",
                                "## Images @ testbed-node-2",
                                "",
                                "+ echo",
                                "+ echo '## Images @ testbed-node-2'",
                                "+ echo",
                                "+ osism container testbed-node-2 images",
                                "REPOSITORY                                                    TAG       IMAGE ID       CREATED        SIZE",
                                "registry.osism.tech/osism/ceph-daemon                         reef      60d930974fb1   22 hours ago   1.27GB",
                                "registry.osism.tech/kolla/keepalived                          2024.2    b2b59a47e345   23 hours ago   286MB",
                                "registry.osism.tech/kolla/grafana                             2024.2    25f6c06d0021   23 hours ago   1.04GB",
                                "registry.osism.tech/kolla/kolla-toolbox                       2024.2    2ff405629e62   23 hours ago   677MB",
                                "registry.osism.tech/kolla/cron                                2024.2    23ec8544c793   23 hours ago   275MB",
                                "registry.osism.tech/kolla/opensearch-dashboards               2024.2    687dcde2bcb8   23 hours ago   1.54GB",
                                "registry.osism.tech/kolla/opensearch                          2024.2    b409ad33e488   23 hours ago   1.57GB",
                                "registry.osism.tech/kolla/memcached                           2024.2    ed2bff2a6973   23 hours ago   276MB",
                                "registry.osism.tech/kolla/proxysql                            2024.2    b4d6db8fd341   24 hours ago   426MB",
                                "registry.osism.tech/kolla/fluentd                             2024.2    851a383a0141   24 hours ago   589MB",
                                "registry.osism.tech/kolla/rabbitmq                            2024.2    a4ab33fd70fd   24 hours ago   332MB",
                                "registry.osism.tech/kolla/haproxy                             2024.2    3fa2a10107a8   24 hours ago   284MB",
                                "registry.osism.tech/kolla/redis-sentinel                      2024.2    5f4f04f4383c   24 hours ago   282MB",
                                "registry.osism.tech/kolla/redis                               2024.2    533c487568ee   24 hours ago   282MB",
                                "registry.osism.tech/kolla/mariadb-server                      2024.2    297f8bb339a1   24 hours ago   462MB",
                                "registry.osism.tech/kolla/prometheus-memcached-exporter       2024.2    dd336c43b779   24 hours ago   308MB",
                                "registry.osism.tech/kolla/prometheus-node-exporter            2024.2    e9fbd7018fbf   24 hours ago   315MB",
                                "registry.osism.tech/kolla/prometheus-mysqld-exporter          2024.2    548293489d76   24 hours ago   310MB",
                                "registry.osism.tech/kolla/prometheus-cadvisor                 2024.2    1520e18eb877   24 hours ago   367MB",
                                "registry.osism.tech/kolla/prometheus-elasticsearch-exporter   2024.2    5b796764074a   24 hours ago   301MB",
                                "registry.osism.tech/kolla/horizon                             2024.2    50eeb61edd36   24 hours ago   1.16GB",
                                "registry.osism.tech/kolla/openvswitch-vswitchd                2024.2    dbcb0c63d3a2   24 hours ago   288MB",
                                "registry.osism.tech/kolla/openvswitch-db-server               2024.2    a8f2511eff1f   24 hours ago   288MB",
                                "registry.osism.tech/kolla/neutron-server                      2024.2    7c6811afdb1b   24 hours ago   1.17GB",
                                "registry.osism.tech/kolla/octavia-api                         2024.2    eaf97bcdf081   24 hours ago   1.06GB",
                                "registry.osism.tech/kolla/octavia-driver-agent                2024.2    4c4abbcba868   24 hours ago   1.06GB",
                                "registry.osism.tech/kolla/octavia-worker                      2024.2    ae8999de916c   24 hours ago   1.04GB",
                                "registry.osism.tech/kolla/octavia-housekeeping                2024.2    e15b03944567   24 hours ago   1.04GB",
                                "registry.osism.tech/kolla/octavia-health-manager              2024.2    41dcecaeaf89   24 hours ago   1.04GB",
                                "registry.osism.tech/kolla/keystone                            2024.2    a6b20f41325d   24 hours ago   1.08GB",
                                "registry.osism.tech/kolla/keystone-fernet                     2024.2    15aaf5c48263   24 hours ago   1.05GB",
                                "registry.osism.tech/kolla/keystone-ssh                        2024.2    2634cde67cfd   24 hours ago   1.05GB",
                                "registry.osism.tech/kolla/cinder-backup                       2024.2    5f8e22fdae27   24 hours ago   1.42GB",
                                "registry.osism.tech/kolla/cinder-volume                       2024.2    29801e980d03   24 hours ago   1.73GB",
                                "registry.osism.tech/kolla/cinder-scheduler                    2024.2    7d98f6182a5d   24 hours ago   1.41GB",
                                "registry.osism.tech/kolla/cinder-api                          2024.2    c16461b7e25c   24 hours ago   1.41GB",
                                "registry.osism.tech/kolla/barbican-worker                     2024.2    cbe9a09e642e   24 hours ago   1GB",
                                "registry.osism.tech/kolla/barbican-keystone-listener          2024.2    8b79c739c5aa   24 hours ago   1GB",
                                "registry.osism.tech/kolla/barbican-api                        2024.2    ec75be03f45f   24 hours ago   1e+03MB",
                                "registry.osism.tech/kolla/nova-scheduler                      2024.2    d13b758badcb   24 hours ago   1.22GB",
                                "registry.osism.tech/kolla/nova-api                            2024.2    fa74d9e1d80c   24 hours ago   1.22GB",
                                "registry.osism.tech/kolla/nova-novncproxy                     2024.2    b0f517fc2110   24 hours ago   1.38GB",
                                "registry.osism.tech/kolla/nova-conductor                      2024.2    cc5099f460a9   24 hours ago   1.22GB",
                                "registry.osism.tech/kolla/glance-api                          2024.2    0c1d9d7160de   24 hours ago   1.11GB",
                                "registry.osism.tech/kolla/magnum-conductor                    2024.2    e4d602a3ad04   24 hours ago   1.25GB",
                                "registry.osism.tech/kolla/magnum-api                          2024.2    083abdbaa2fa   24 hours ago   1.14GB",
                                "registry.osism.tech/kolla/placement-api                       2024.2    dfc171655fae   24 hours ago   985MB",
                                "registry.osism.tech/kolla/designate-worker                    2024.2    e9363cf6baa5   24 hours ago   998MB",
                                "registry.osism.tech/kolla/designate-producer                  2024.2    1d3f00ae0cb0   24 hours ago   994MB",
                                "registry.osism.tech/kolla/designate-mdns                      2024.2    b74dea85ff31   24 hours ago   994MB",
                                "registry.osism.tech/kolla/designate-backend-bind9             2024.2    71020f79d0b8   24 hours ago   998MB",
                                "registry.osism.tech/kolla/designate-central                   2024.2    e270fdf55dd9   24 hours ago   993MB",
                                "registry.osism.tech/kolla/designate-api                       2024.2    2b71cfe02281   24 hours ago   994MB",
                                "registry.osism.tech/kolla/ovn-nb-db-server                    2024.2    91492a927f10   24 hours ago   850MB",
                                "registry.osism.tech/kolla/ovn-sb-db-server                    2024.2    c3d787fcf696   24 hours ago   850MB",
                                "registry.osism.tech/kolla/ovn-controller                      2024.2    df4830242254   24 hours ago   850MB",
                                "registry.osism.tech/kolla/ovn-northd                          2024.2    521cb38ac60e   24 hours ago   850MB",
                                "+ sh -c /opt/configuration/scripts/check-services.sh",
                                "+ set -e",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=true",
                                "++ TEMPEST=true",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]",
                                "+ sh -c /opt/configuration/scripts/check/100-ceph-with-ansible.sh",
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/configuration/scripts/manager-version.sh",
                                "+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml",
                                "",
                                "# Ceph status",
                                "",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "+ echo",
                                "+ echo '# Ceph status'",
                                "+ echo",
                                "+ ceph -s",
                                "  cluster:",
                                "    id:     11111111-1111-1111-1111-111111111111",
                                "    health: HEALTH_OK",
                                " ",
                                "  services:",
                                "    mon: 3 daemons, quorum testbed-node-0,testbed-node-1,testbed-node-2 (age 26m)",
                                "    mgr: testbed-node-1(active, since 14m), standbys: testbed-node-0, testbed-node-2",
                                "    mds: 1/1 daemons up, 2 standby",
                                "    osd: 6 osds: 6 up (since 22m), 6 in (since 23m)",
                                "    rgw: 3 daemons active (3 hosts, 1 zones)",
                                " ",
                                "  data:",
                                "    volumes: 1/1 healthy",
                                "    pools:   14 pools, 401 pgs",
                                "    objects: 556 objects, 2.2 GiB",
                                "    usage:   7.1 GiB used, 113 GiB / 120 GiB avail",
                                "    pgs:     401 active+clean",
                                " ",
                                "",
                                "# Ceph versions",
                                "",
                                "+ echo",
                                "+ echo '# Ceph versions'",
                                "+ echo",
                                "+ ceph versions",
                                "{",
                                "    \"mon\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3",
                                "    },",
                                "    \"mgr\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3",
                                "    },",
                                "    \"osd\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 6",
                                "    },",
                                "    \"mds\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3",
                                "    },",
                                "    \"rgw\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3",
                                "    },",
                                "    \"overall\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 18",
                                "    }",
                                "}",
                                "",
                                "# Ceph OSD tree",
                                "",
                                "+ echo",
                                "+ echo '# Ceph OSD tree'",
                                "+ echo",
                                "+ ceph osd df tree",
                                "ID  CLASS  WEIGHT   REWEIGHT  SIZE     RAW USE   DATA     OMAP     META     AVAIL    %USE  VAR   PGS  STATUS  TYPE NAME              ",
                                "-1         0.11691         -  120 GiB   7.1 GiB  6.7 GiB    6 KiB  421 MiB  113 GiB  5.91  1.00    -          root default           ",
                                "-3         0.03897         -   40 GiB   2.4 GiB  2.2 GiB    2 KiB  139 MiB   38 GiB  5.91  1.00    -              host testbed-node-3",
                                " 0    hdd  0.01949   1.00000   20 GiB   1.0 GiB  955 MiB    1 KiB   70 MiB   19 GiB  5.01  0.85  189      up          osd.0          ",
                                " 3    hdd  0.01949   1.00000   20 GiB   1.4 GiB  1.3 GiB    1 KiB   70 MiB   19 GiB  6.81  1.15  201      up          osd.3          ",
                                "-5         0.03897         -   40 GiB   2.4 GiB  2.2 GiB    2 KiB  143 MiB   38 GiB  5.92  1.00    -              host testbed-node-4",
                                " 1    hdd  0.01949   1.00000   20 GiB  1009 MiB  940 MiB    1 KiB   70 MiB   19 GiB  4.93  0.83  195      up          osd.1          ",
                                " 5    hdd  0.01949   1.00000   20 GiB   1.4 GiB  1.3 GiB    1 KiB   74 MiB   19 GiB  6.90  1.17  197      up          osd.5          ",
                                "-7         0.03897         -   40 GiB   2.4 GiB  2.2 GiB    2 KiB  139 MiB   38 GiB  5.91  1.00    -              host testbed-node-5",
                                " 2    hdd  0.01949   1.00000   20 GiB   1.6 GiB  1.5 GiB    1 KiB   70 MiB   18 GiB  7.79  1.32  198      up          osd.2          ",
                                " 4    hdd  0.01949   1.00000   20 GiB   825 MiB  755 MiB    1 KiB   70 MiB   19 GiB  4.03  0.68  190      up          osd.4          ",
                                "                       TOTAL  120 GiB   7.1 GiB  6.7 GiB  9.3 KiB  421 MiB  113 GiB  5.91                                            ",
                                "MIN/MAX VAR: 0.68/1.32  STDDEV: 1.33",
                                "",
                                "# Ceph monitor status",
                                "",
                                "+ echo",
                                "+ echo '# Ceph monitor status'",
                                "+ echo",
                                "+ ceph mon stat",
                                "e1: 3 mons at {testbed-node-0=[v2:192.168.16.10:3300/0,v1:192.168.16.10:6789/0],testbed-node-1=[v2:192.168.16.11:3300/0,v1:192.168.16.11:6789/0],testbed-node-2=[v2:192.168.16.12:3300/0,v1:192.168.16.12:6789/0]} removed_ranks: {} disallowed_leaders: {}, election epoch 8, leader 0 testbed-node-0, quorum 0,1,2 testbed-node-0,testbed-node-1,testbed-node-2",
                                "",
                                "# Ceph quorum status",
                                "",
                                "+ echo",
                                "+ echo '# Ceph quorum status'",
                                "+ echo",
                                "+ jq",
                                "+ ceph quorum_status",
                                "{",
                                "  \"election_epoch\": 8,",
                                "  \"quorum\": [",
                                "    0,",
                                "    1,",
                                "    2",
                                "  ],",
                                "  \"quorum_names\": [",
                                "    \"testbed-node-0\",",
                                "    \"testbed-node-1\",",
                                "    \"testbed-node-2\"",
                                "  ],",
                                "  \"quorum_leader_name\": \"testbed-node-0\",",
                                "  \"quorum_age\": 1582,",
                                "  \"features\": {",
                                "    \"quorum_con\": \"4540138322906710015\",",
                                "    \"quorum_mon\": [",
                                "      \"kraken\",",
                                "      \"luminous\",",
                                "      \"mimic\",",
                                "      \"osdmap-prune\",",
                                "      \"nautilus\",",
                                "      \"octopus\",",
                                "      \"pacific\",",
                                "      \"elector-pinging\",",
                                "      \"quincy\",",
                                "      \"reef\"",
                                "    ]",
                                "  },",
                                "  \"monmap\": {",
                                "    \"epoch\": 1,",
                                "    \"fsid\": \"11111111-1111-1111-1111-111111111111\",",
                                "    \"modified\": \"2026-03-21T00:48:46.754580Z\",",
                                "    \"created\": \"2026-03-21T00:48:46.754580Z\",",
                                "    \"min_mon_release\": 18,",
                                "    \"min_mon_release_name\": \"reef\",",
                                "    \"election_strategy\": 1,",
                                "    \"disallowed_leaders: \": \"\",",
                                "    \"stretch_mode\": false,",
                                "    \"tiebreaker_mon\": \"\",",
                                "    \"removed_ranks: \": \"\",",
                                "    \"features\": {",
                                "      \"persistent\": [",
                                "        \"kraken\",",
                                "        \"luminous\",",
                                "        \"mimic\",",
                                "        \"osdmap-prune\",",
                                "        \"nautilus\",",
                                "        \"octopus\",",
                                "        \"pacific\",",
                                "        \"elector-pinging\",",
                                "        \"quincy\",",
                                "        \"reef\"",
                                "      ],",
                                "      \"optional\": []",
                                "    },",
                                "    \"mons\": [",
                                "      {",
                                "        \"rank\": 0,",
                                "        \"name\": \"testbed-node-0\",",
                                "        \"public_addrs\": {",
                                "          \"addrvec\": [",
                                "            {",
                                "              \"type\": \"v2\",",
                                "              \"addr\": \"192.168.16.10:3300\",",
                                "              \"nonce\": 0",
                                "            },",
                                "            {",
                                "              \"type\": \"v1\",",
                                "              \"addr\": \"192.168.16.10:6789\",",
                                "              \"nonce\": 0",
                                "            }",
                                "          ]",
                                "        },",
                                "        \"addr\": \"192.168.16.10:6789/0\",",
                                "        \"public_addr\": \"192.168.16.10:6789/0\",",
                                "        \"priority\": 0,",
                                "        \"weight\": 0,",
                                "        \"crush_location\": \"{}\"",
                                "      },",
                                "      {",
                                "        \"rank\": 1,",
                                "        \"name\": \"testbed-node-1\",",
                                "        \"public_addrs\": {",
                                "          \"addrvec\": [",
                                "            {",
                                "              \"type\": \"v2\",",
                                "              \"addr\": \"192.168.16.11:3300\",",
                                "              \"nonce\": 0",
                                "            },",
                                "            {",
                                "              \"type\": \"v1\",",
                                "              \"addr\": \"192.168.16.11:6789\",",
                                "              \"nonce\": 0",
                                "            }",
                                "          ]",
                                "        },",
                                "        \"addr\": \"192.168.16.11:6789/0\",",
                                "        \"public_addr\": \"192.168.16.11:6789/0\",",
                                "        \"priority\": 0,",
                                "        \"weight\": 0,",
                                "        \"crush_location\": \"{}\"",
                                "      },",
                                "      {",
                                "        \"rank\": 2,",
                                "        \"name\": \"testbed-node-2\",",
                                "        \"public_addrs\": {",
                                "          \"addrvec\": [",
                                "            {",
                                "              \"type\": \"v2\",",
                                "              \"addr\": \"192.168.16.12:3300\",",
                                "              \"nonce\": 0",
                                "            },",
                                "            {",
                                "              \"type\": \"v1\",",
                                "              \"addr\": \"192.168.16.12:6789\",",
                                "              \"nonce\": 0",
                                "            }",
                                "          ]",
                                "        },",
                                "        \"addr\": \"192.168.16.12:6789/0\",",
                                "        \"public_addr\": \"192.168.16.12:6789/0\",",
                                "        \"priority\": 0,",
                                "        \"weight\": 0,",
                                "        \"crush_location\": \"{}\"",
                                "      }",
                                "    ]",
                                "  }",
                                "}",
                                "",
                                "# Ceph free space status",
                                "",
                                "+ echo",
                                "+ echo '# Ceph free space status'",
                                "+ echo",
                                "+ ceph df",
                                "--- RAW STORAGE ---",
                                "CLASS     SIZE    AVAIL     USED  RAW USED  %RAW USED",
                                "hdd    120 GiB  113 GiB  7.1 GiB   7.1 GiB       5.91",
                                "TOTAL  120 GiB  113 GiB  7.1 GiB   7.1 GiB       5.91",
                                " ",
                                "--- POOLS ---",
                                "POOL                       ID  PGS   STORED  OBJECTS     USED  %USED  MAX AVAIL",
                                ".mgr                        1    1  577 KiB        2  1.1 MiB      0     52 GiB",
                                "cephfs_data                 2   32      0 B        0      0 B      0     35 GiB",
                                "cephfs_metadata             3   16  4.4 KiB       22   96 KiB      0     35 GiB",
                                "default.rgw.buckets.data    4   32      0 B        0      0 B      0     35 GiB",
                                "default.rgw.buckets.index   5   32      0 B        0      0 B      0     35 GiB",
                                "default.rgw.control         6   32      0 B        8      0 B      0     35 GiB",
                                "default.rgw.log             7   32  3.6 KiB      209  408 KiB      0     35 GiB",
                                "default.rgw.meta            8   32      0 B        0      0 B      0     35 GiB",
                                ".rgw.root                   9   32  3.9 KiB        8   64 KiB      0     52 GiB",
                                "backups                    10   32     19 B        2   12 KiB      0     35 GiB",
                                "volumes                    11   32     19 B        2   12 KiB      0     35 GiB",
                                "images                     12   32  2.2 GiB      299  6.7 GiB   5.99     35 GiB",
                                "metrics                    13   32     19 B        2   12 KiB      0     35 GiB",
                                "vms                        14   32     19 B        2   12 KiB      0     35 GiB",
                                "++ semver latest 5.0.0",
                                "+ [[ -1 -eq -1 ]]",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ [[ ! -e /etc/redhat-release ]]",
                                "+ osism apply facts",
                                "\u001b[32m2026-03-21 01:15:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of facts.\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 05981b19-0a4e-4355-8129-9611087a39d0 (facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 01:15:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 05981b19-0a4e-4355-8129-9611087a39d0 (facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role facts] ********************************************************",
                                "",
                                "TASK [osism.commons.facts : Create custom facts directory] *********************",
                                "Saturday 21 March 2026  01:15:48 +0000 (0:00:00.347)       0:00:00.347 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Copy fact files] ***********************************",
                                "Saturday 21 March 2026  01:15:49 +0000 (0:00:01.413)       0:00:01.761 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Saturday 21 March 2026  01:15:50 +0000 (0:00:01.325)       0:00:03.086 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts if using --limit] *****************************",
                                "",
                                "TASK [Gather facts for all hosts] **********************************************",
                                "Saturday 21 March 2026  01:15:55 +0000 (0:00:04.212)       0:00:07.298 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  01:15:55 +0000 (0:00:00.705)       0:00:08.004 ******** ",
                                "=============================================================================== ",
                                "Gathers facts about hosts ----------------------------------------------- 4.21s",
                                "osism.commons.facts : Create custom facts directory --------------------- 1.41s",
                                "osism.commons.facts : Copy fact files ----------------------------------- 1.33s",
                                "Gather facts for all hosts ---------------------------------------------- 0.71s",
                                "+ osism validate ceph-mons",
                                "",
                                "PLAY [Ceph validate mons] ******************************************************",
                                "",
                                "TASK [Get timestamp for report file] *******************************************",
                                "Saturday 21 March 2026  01:16:11 +0000 (0:00:00.540)       0:00:00.541 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Create report output directory] ******************************************",
                                "Saturday 21 March 2026  01:16:12 +0000 (0:00:00.988)       0:00:01.529 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Define report vars] ******************************************************",
                                "Saturday 21 March 2026  01:16:12 +0000 (0:00:00.692)       0:00:02.221 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Prepare test data for container existance test] **************************",
                                "Saturday 21 March 2026  01:16:12 +0000 (0:00:00.135)       0:00:02.357 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Get container info] ******************************************************",
                                "Saturday 21 March 2026  01:16:13 +0000 (0:00:00.284)       0:00:02.642 ******** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if container is missing] ***********************",
                                "Saturday 21 March 2026  01:16:14 +0000 (0:00:01.359)       0:00:04.002 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if container is existing] **********************",
                                "Saturday 21 March 2026  01:16:14 +0000 (0:00:00.279)       0:00:04.282 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Saturday 21 March 2026  01:16:15 +0000 (0:00:00.326)       0:00:04.608 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if ceph-mon is not running] ********************",
                                "Saturday 21 March 2026  01:16:15 +0000 (0:00:00.309)       0:00:04.918 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if ceph-mon is running] ************************",
                                "Saturday 21 March 2026  01:16:15 +0000 (0:00:00.435)       0:00:05.353 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Saturday 21 March 2026  01:16:16 +0000 (0:00:00.297)       0:00:05.651 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Saturday 21 March 2026  01:16:16 +0000 (0:00:00.239)       0:00:05.891 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Saturday 21 March 2026  01:16:16 +0000 (0:00:00.241)       0:00:06.132 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Saturday 21 March 2026  01:16:16 +0000 (0:00:00.238)       0:00:06.371 ******** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Saturday 21 March 2026  01:16:17 +0000 (0:00:00.081)       0:00:06.452 ******** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Saturday 21 March 2026  01:16:17 +0000 (0:00:00.069)       0:00:06.522 ******** ",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Saturday 21 March 2026  01:16:17 +0000 (0:00:00.216)       0:00:06.738 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail due to missing containers] ******************************************",
                                "Saturday 21 March 2026  01:16:17 +0000 (0:00:00.254)       0:00:06.993 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Prepare quorum test vars] ************************************************",
                                "Saturday 21 March 2026  01:16:17 +0000 (0:00:00.234)       0:00:07.228 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Get monmap info from one mon container] **********************************",
                                "Saturday 21 March 2026  01:16:17 +0000 (0:00:00.128)       0:00:07.357 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set quorum test data] ****************************************************",
                                "Saturday 21 March 2026  01:16:19 +0000 (0:00:01.946)       0:00:09.303 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail quorum test if not all monitors are in quorum] **********************",
                                "Saturday 21 March 2026  01:16:20 +0000 (0:00:00.301)       0:00:09.605 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass quorum test if all monitors are in quorum] **************************",
                                "Saturday 21 March 2026  01:16:20 +0000 (0:00:00.131)       0:00:09.737 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set fsid test vars] ******************************************************",
                                "Saturday 21 March 2026  01:16:20 +0000 (0:00:00.330)       0:00:10.067 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail Cluster FSID test if FSID does not match configuration] *************",
                                "Saturday 21 March 2026  01:16:20 +0000 (0:00:00.285)       0:00:10.353 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass Cluster FSID test if it matches configuration] **********************",
                                "Saturday 21 March 2026  01:16:21 +0000 (0:00:00.125)       0:00:10.479 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Prepare status test vars] ************************************************",
                                "Saturday 21 March 2026  01:16:21 +0000 (0:00:00.124)       0:00:10.604 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Gather status data] ******************************************************",
                                "Saturday 21 March 2026  01:16:21 +0000 (0:00:00.288)       0:00:10.892 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set health test data] ****************************************************",
                                "Saturday 21 March 2026  01:16:23 +0000 (0:00:01.556)       0:00:12.449 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail cluster-health if health is not acceptable] *************************",
                                "Saturday 21 March 2026  01:16:23 +0000 (0:00:00.318)       0:00:12.768 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass cluster-health if health is acceptable] *****************************",
                                "Saturday 21 March 2026  01:16:23 +0000 (0:00:00.132)       0:00:12.901 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail cluster-health if health is not acceptable (strict)] ****************",
                                "Saturday 21 March 2026  01:16:23 +0000 (0:00:00.142)       0:00:13.044 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass cluster-health if status is OK (strict)] ****************************",
                                "Saturday 21 March 2026  01:16:23 +0000 (0:00:00.144)       0:00:13.188 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set validation result to passed if no test failed] ***********************",
                                "Saturday 21 March 2026  01:16:23 +0000 (0:00:00.135)       0:00:13.323 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Set validation result to failed if a test failed] ************************",
                                "Saturday 21 March 2026  01:16:24 +0000 (0:00:00.260)       0:00:13.584 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Saturday 21 March 2026  01:16:24 +0000 (0:00:00.247)       0:00:13.832 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Saturday 21 March 2026  01:16:26 +0000 (0:00:01.746)       0:00:15.578 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Saturday 21 March 2026  01:16:26 +0000 (0:00:00.250)       0:00:15.829 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Saturday 21 March 2026  01:16:27 +0000 (0:00:00.645)       0:00:16.474 ******** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Saturday 21 March 2026  01:16:27 +0000 (0:00:00.070)       0:00:16.545 ******** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Saturday 21 March 2026  01:16:27 +0000 (0:00:00.076)       0:00:16.621 ******** ",
                                "",
                                "RUNNING HANDLER [Write report file] ********************************************",
                                "Saturday 21 March 2026  01:16:27 +0000 (0:00:00.076)       0:00:16.698 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Saturday 21 March 2026  01:16:28 +0000 (0:00:01.248)       0:00:17.947 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": [\u001b[0m",
                                "\u001b[0;32m        \"Validator run completed.\",\u001b[0m",
                                "\u001b[0;32m        \"You can find the report file here:\",\u001b[0m",
                                "\u001b[0;32m        \"/opt/reports/validator/ceph-mons-validator-2026-03-21T01:16:12+00:00-report.json\",\u001b[0m",
                                "\u001b[0;32m        \"on the following host:\",\u001b[0m",
                                "\u001b[0;32m        \"testbed-manager\"\u001b[0m",
                                "\u001b[0;32m    ]\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=13  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  01:16:28 +0000 (0:00:00.407)       0:00:18.354 ******** ",
                                "=============================================================================== ",
                                "Get monmap info from one mon container ---------------------------------- 1.95s",
                                "Aggregate test results step one ----------------------------------------- 1.75s",
                                "Gather status data ------------------------------------------------------ 1.56s",
                                "Get container info ------------------------------------------------------ 1.36s",
                                "Write report file ------------------------------------------------------- 1.25s",
                                "Get timestamp for report file ------------------------------------------- 0.99s",
                                "Create report output directory ------------------------------------------ 0.69s",
                                "Aggregate test results step three --------------------------------------- 0.65s",
                                "Set test result to failed if ceph-mon is not running -------------------- 0.44s",
                                "Print report file information ------------------------------------------- 0.41s",
                                "Flush handlers ---------------------------------------------------------- 0.37s",
                                "Pass quorum test if all monitors are in quorum -------------------------- 0.33s",
                                "Set test result to passed if container is existing ---------------------- 0.33s",
                                "Set health test data ---------------------------------------------------- 0.32s",
                                "Prepare test data ------------------------------------------------------- 0.31s",
                                "Set quorum test data ---------------------------------------------------- 0.30s",
                                "Set test result to passed if ceph-mon is running ------------------------ 0.30s",
                                "Prepare status test vars ------------------------------------------------ 0.29s",
                                "Set fsid test vars ------------------------------------------------------ 0.29s",
                                "Prepare test data for container existance test -------------------------- 0.28s",
                                "+ osism validate ceph-mgrs",
                                "",
                                "PLAY [Ceph validate mgrs] ******************************************************",
                                "",
                                "TASK [Get timestamp for report file] *******************************************",
                                "Saturday 21 March 2026  01:16:44 +0000 (0:00:00.506)       0:00:00.506 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Create report output directory] ******************************************",
                                "Saturday 21 March 2026  01:16:45 +0000 (0:00:00.986)       0:00:01.492 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Define report vars] ******************************************************",
                                "Saturday 21 March 2026  01:16:45 +0000 (0:00:00.688)       0:00:02.181 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Prepare test data for container existance test] **************************",
                                "Saturday 21 March 2026  01:16:46 +0000 (0:00:00.126)       0:00:02.308 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Get container info] ******************************************************",
                                "Saturday 21 March 2026  01:16:46 +0000 (0:00:00.276)       0:00:02.585 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if container is missing] ***********************",
                                "Saturday 21 March 2026  01:16:47 +0000 (0:00:01.382)       0:00:03.967 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if container is existing] **********************",
                                "Saturday 21 March 2026  01:16:48 +0000 (0:00:00.287)       0:00:04.254 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Saturday 21 March 2026  01:16:48 +0000 (0:00:00.300)       0:00:04.554 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if ceph-mgr is not running] ********************",
                                "Saturday 21 March 2026  01:16:48 +0000 (0:00:00.300)       0:00:04.855 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if ceph-mgr is running] ************************",
                                "Saturday 21 March 2026  01:16:49 +0000 (0:00:00.457)       0:00:05.312 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Saturday 21 March 2026  01:16:49 +0000 (0:00:00.290)       0:00:05.602 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Saturday 21 March 2026  01:16:49 +0000 (0:00:00.233)       0:00:05.835 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Saturday 21 March 2026  01:16:49 +0000 (0:00:00.262)       0:00:06.098 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Saturday 21 March 2026  01:16:50 +0000 (0:00:00.237)       0:00:06.336 ******** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Saturday 21 March 2026  01:16:50 +0000 (0:00:00.094)       0:00:06.430 ******** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Saturday 21 March 2026  01:16:50 +0000 (0:00:00.076)       0:00:06.507 ******** ",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Saturday 21 March 2026  01:16:50 +0000 (0:00:00.235)       0:00:06.743 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail due to missing containers] ******************************************",
                                "Saturday 21 March 2026  01:16:50 +0000 (0:00:00.262)       0:00:07.006 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Define mgr module test vars] *********************************************",
                                "Saturday 21 March 2026  01:16:51 +0000 (0:00:00.271)       0:00:07.278 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Gather list of mgr modules] **********************************************",
                                "Saturday 21 March 2026  01:16:51 +0000 (0:00:00.132)       0:00:07.410 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Parse mgr module list from json] *****************************************",
                                "Saturday 21 March 2026  01:16:53 +0000 (0:00:02.099)       0:00:09.510 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Extract list of enabled mgr modules] *************************************",
                                "Saturday 21 March 2026  01:16:53 +0000 (0:00:00.240)       0:00:09.751 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail test if mgr modules are disabled that should be enabled] ************",
                                "Saturday 21 March 2026  01:16:53 +0000 (0:00:00.289)       0:00:10.041 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass test if required mgr modules are enabled] ***************************",
                                "Saturday 21 March 2026  01:16:53 +0000 (0:00:00.114)       0:00:10.155 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set validation result to passed if no test failed] ***********************",
                                "Saturday 21 March 2026  01:16:54 +0000 (0:00:00.138)       0:00:10.294 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Set validation result to failed if a test failed] ************************",
                                "Saturday 21 March 2026  01:16:54 +0000 (0:00:00.250)       0:00:10.544 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Saturday 21 March 2026  01:16:54 +0000 (0:00:00.239)       0:00:10.784 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Saturday 21 March 2026  01:16:56 +0000 (0:00:01.475)       0:00:12.260 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Saturday 21 March 2026  01:16:56 +0000 (0:00:00.268)       0:00:12.528 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Saturday 21 March 2026  01:16:56 +0000 (0:00:00.263)       0:00:12.792 ******** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Saturday 21 March 2026  01:16:56 +0000 (0:00:00.074)       0:00:12.866 ******** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Saturday 21 March 2026  01:16:56 +0000 (0:00:00.069)       0:00:12.936 ******** ",
                                "",
                                "RUNNING HANDLER [Write report file] ********************************************",
                                "Saturday 21 March 2026  01:16:56 +0000 (0:00:00.078)       0:00:13.014 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Saturday 21 March 2026  01:16:58 +0000 (0:00:01.327)       0:00:14.342 ******** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": [\u001b[0m",
                                "\u001b[0;32m        \"Validator run completed.\",\u001b[0m",
                                "\u001b[0;32m        \"You can find the report file here:\",\u001b[0m",
                                "\u001b[0;32m        \"/opt/reports/validator/ceph-mgrs-validator-2026-03-21T01:16:45+00:00-report.json\",\u001b[0m",
                                "\u001b[0;32m        \"on the following host:\",\u001b[0m",
                                "\u001b[0;32m        \"testbed-manager\"\u001b[0m",
                                "\u001b[0;32m    ]\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  01:16:58 +0000 (0:00:00.392)       0:00:14.734 ******** ",
                                "=============================================================================== ",
                                "Gather list of mgr modules ---------------------------------------------- 2.10s",
                                "Aggregate test results step one ----------------------------------------- 1.48s",
                                "Get container info ------------------------------------------------------ 1.38s",
                                "Write report file ------------------------------------------------------- 1.33s",
                                "Get timestamp for report file ------------------------------------------- 0.99s",
                                "Create report output directory ------------------------------------------ 0.69s",
                                "Set test result to failed if ceph-mgr is not running -------------------- 0.46s",
                                "Flush handlers ---------------------------------------------------------- 0.41s",
                                "Print report file information ------------------------------------------- 0.39s",
                                "Prepare test data ------------------------------------------------------- 0.30s",
                                "Set test result to passed if container is existing ---------------------- 0.30s",
                                "Set test result to passed if ceph-mgr is running ------------------------ 0.29s",
                                "Extract list of enabled mgr modules ------------------------------------- 0.29s",
                                "Set test result to failed if container is missing ----------------------- 0.29s",
                                "Prepare test data for container existance test -------------------------- 0.28s",
                                "Fail due to missing containers ------------------------------------------ 0.27s",
                                "Aggregate test results step two ----------------------------------------- 0.27s",
                                "Aggregate test results step three --------------------------------------- 0.26s",
                                "Print report file information ------------------------------------------- 0.26s",
                                "Aggregate test results step two ----------------------------------------- 0.26s",
                                "+ osism validate ceph-osds",
                                "",
                                "PLAY [Ceph validate OSDs] ******************************************************",
                                "",
                                "TASK [Get timestamp for report file] *******************************************",
                                "Saturday 21 March 2026  01:17:14 +0000 (0:00:00.497)       0:00:00.498 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Saturday 21 March 2026  01:17:14 +0000 (0:00:00.949)       0:00:01.447 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Create report output directory] ******************************************",
                                "Saturday 21 March 2026  01:17:15 +0000 (0:00:00.238)       0:00:01.686 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Define report vars] ******************************************************",
                                "Saturday 21 March 2026  01:17:15 +0000 (0:00:00.672)       0:00:02.358 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Define OSD test variables] ***********************************************",
                                "Saturday 21 March 2026  01:17:15 +0000 (0:00:00.123)       0:00:02.482 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate OSD devices for each host] *************************************",
                                "Saturday 21 March 2026  01:17:16 +0000 (0:00:00.123)       0:00:02.605 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Define OSD test variables] ***********************************************",
                                "Saturday 21 March 2026  01:17:16 +0000 (0:00:00.450)       0:00:03.056 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate OSD devices for each host] *************************************",
                                "Saturday 21 March 2026  01:17:16 +0000 (0:00:00.150)       0:00:03.206 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate total number of OSDs in cluster] *******************************",
                                "Saturday 21 March 2026  01:17:17 +0000 (0:00:00.356)       0:00:03.563 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Saturday 21 March 2026  01:17:17 +0000 (0:00:00.325)       0:00:03.888 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get list of ceph-osd containers on host] *********************************",
                                "Saturday 21 March 2026  01:17:17 +0000 (0:00:00.287)       0:00:04.176 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'b3834e4b9797d2ff4db9f4f433ff64037dedaa8e4e196486c83b6934b2b415da', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '04611561164cd51fc3a1ed5e9f60065177b9e751a138495a53804363d317566f', 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '686e42457e4a98d64855761766743009124da2d28c2111f4f56ec71ba08a5890', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '8a25dc8033b6564150316723eafb179e510067642a65fe7adc86d9e644ab9b14', 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 11 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '1ded55d684f98646e1b9581d31ebbf95ec0c9c8c87b69d8a1e3f6f77d9269b2a', 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 11 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '5a861065359ee5c8d9fcbe7cdcac9b4666265e7236c1deaa3d14cce10a7ebbb2', 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 12 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '0022450ab44c1664f8d92910e0b3377c16e5cf0f56e88032ea7b190c3d50d895', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 13 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '674e9590eb63fbaa5efe8cde9a89c3538774a40289a43b678d8aae81c688696e', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-rgw-default-testbed-node-3-rgw0', 'state': 'running', 'status': 'Up 20 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'fab7712f40b620b557b418bc8a67d66bfd9cab7ace760248fb3fc076f460282c', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-mds-testbed-node-3', 'state': 'running', 'status': 'Up 21 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'b6044d59fee87068d24f2a15531f68581ffb047be92817620778e740a75cc1c6', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-crash-testbed-node-3', 'state': 'running', 'status': 'Up 22 minutes'}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'id': 'f80f9c86195d357ba685e4352428814027d0fab54202bb85f0680846813787de', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-3', 'state': 'running', 'status': 'Up 23 minutes'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'id': 'ced04366240e3aef9d90dc1cbb93c3ac0c4cee185f775b81dc73e7189c1d1e1c', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-0', 'state': 'running', 'status': 'Up 23 minutes'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '501b7122783ecf7a8d2a9445be6ef667b253ef950b196c910b800740ae2a1b5f', 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 26 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'ea950199405bc692edb293a5b3a5c718909ca184ab5d6a0dfd78b72ecfc51961', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 27 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '318388aef55ec0589d19a19ada3e0135c5c78e1a5f96c2573b650800443d3bc1', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 28 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '0ae70184f8566bb9deeafcbc6473423913b6bda2f6a90e5be8dbe3be275614ee', 'image': 'registry.osism.tech/kolla/cron:2024.2', 'name': '/cron', 'state': 'running', 'status': 'Up 28 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'bbf4cca6cfba90de64cadaa5eefae31d6081e3cf05b17890833266c229965fa5', 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 29 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '3820df04684d6d4bd30e1c4e426bfa66c74d53d8bbb13dd34b07f45716def56e', 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'name': '/fluentd', 'state': 'running', 'status': 'Up 29 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '4bcee744b0035069a448ebf926bbea6186f7d70ca7add1390f48951354d848ea', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '3f5e9d5cb4af813b127ff2510d0a21d593016f3a5e5f2b0255c6434e12546eb8', 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '82bb0e0b64eb13f425b2df3bcd5b458f4bdd35ea0c9092ec6ac83670dd2dbc4f', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '6000eeba069a7193758f7419466fd78c1dd986f8d0afab94b95b9f79577223d1', 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 11 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '1f6f3a4e0dd8c9d0780ae2a323f296b39901d8da51a2baf7fc96999ffe77469f', 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 11 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '33f610c36ac8458fce46755538cae737bc46001eaca9905e6d3f0e097662f9a4', 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 12 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'c691f7d217a9d8449cebbb687dbac850d02b0487365b2ad60434c3f6f589d8c6', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 12 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'f9709df7ba324efdcff3a8cf25d4bd1d1e79c443f1462e90429c0dfb0f219088', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-rgw-default-testbed-node-4-rgw0', 'state': 'running', 'status': 'Up 20 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '474c051b9924801f91e33a2c599e059a380ba100fa39ad81769345b1e9558599', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-mds-testbed-node-4', 'state': 'running', 'status': 'Up 21 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '8425dcc075d1c743c3594a66cc973c80cad5c730ef69eb67ce6327dee0a875a2', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-crash-testbed-node-4', 'state': 'running', 'status': 'Up 22 minutes'}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'id': '9c825b9e0312f119368d2a0a7f03bd3dbf08b583add2062e8abd33e426500564', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-5', 'state': 'running', 'status': 'Up 23 minutes'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'id': '1add874758f001f932b28d061b71a9f39dc0f673ca85f86311e6920ca2db45b2', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-1', 'state': 'running', 'status': 'Up 23 minutes'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '670e4278f5d14145d6995bbc5b4d9aec7fa1d06776d8bc99c90a6e70214071e0', 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 26 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '822bf99a7617765d185192fc3c4880968cae2b43f98e55b2e4d4493d029f1826', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 27 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'd5f3c47c1de714009c246574488dea7b23dc8d2ef9bec0b35ac6032798555398', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 28 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '50aed0d9cdc3614469ef918fa6b903fff4cc2fa1103421c95ab18bb6c2e22c63', 'image': 'registry.osism.tech/kolla/cron:2024.2', 'name': '/cron', 'state': 'running', 'status': 'Up 28 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'ba8abe4b0317466dc1f636a0370fa0dcde2ef017dbd3caccb95ef5d13cc0f8fd', 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 29 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '9502548dd53e8359c6f99f0af769e6c2c54ae43bb11b57fc02653eeec7be811f', 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'name': '/fluentd', 'state': 'running', 'status': 'Up 29 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '4217b11e61a8f60f98ce804bc5bb30eb314cd4304461f7b63f1089f799acbb45', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '2b0e43585d8764f9c5c739ef4588b6b9ae88abcb577d0fed777c3508e40623db', 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '5e9ba54c5522ef2df374bab9292de1c7df20874ea099b1331c00ab0d9c062b80', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '1b8a7e9e094d41c11fd6bd46552a4366960e7ad6098a978d1310bb9a81fd1abb', 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 11 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '9b24a1153e98cbb631a896c63e9ee756923c22be26f175f91cc5dd8206f86797', 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 11 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '19601c0a2d6c2b10f8e644aebb2f1048678f0b6c0a8c037870184fc5ab9521a8', 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 12 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'b439d9b1776dcd4a9e555ce5a0065c128b6a9494afe441076937c4fac164db68', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 12 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '93a749a896acce9595ef8c691f29ca9095d7880856f41bfe32b6ca3286d598fa', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-rgw-default-testbed-node-5-rgw0', 'state': 'running', 'status': 'Up 20 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'fce8668785c102a8ae343c034b76ce0ee6f4155aaaff792ec316dd5fc935062f', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-mds-testbed-node-5', 'state': 'running', 'status': 'Up 21 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'c7183ffbca2bbe4fa711520a7c4f36a61fa21ca24441d0bceb49221e7d22077d', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-crash-testbed-node-5', 'state': 'running', 'status': 'Up 22 minutes'}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'id': 'b39b31c03daf16fdb12e36656287c8ae5d75f5f32dec390e1a6b6fa6d8bf7a94', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-2', 'state': 'running', 'status': 'Up 23 minutes'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'id': '536e57de3f664f37f632d6d51267d8ead62a6fb4d1a06749969ddb6edf3119f4', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-4', 'state': 'running', 'status': 'Up 23 minutes'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '21301f5c0879a6affc5198ad889a1eb895d48ae7f32f3404a8e266c86d0be63b', 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 26 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '7c2076a2645cbf1f4568a36b52457a4b644600723a28afb1c29f09f5fece12c8', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 27 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '6696533b70d60f4d778334f6c795e1a3752c38b63ed2e20c33d28137394ebd73', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 28 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '15cc45b98c17825e6ac3d61d38e3cb9038433e22d41bba2f63922612822b9763', 'image': 'registry.osism.tech/kolla/cron:2024.2', 'name': '/cron', 'state': 'running', 'status': 'Up 28 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '4e853a7f8e10afbcc48c23e687d54fc607d86a9a6186d93057c82cac9c869f38', 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 29 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '5913715c59140ecf4be6c29f797c050fb5c792a82330d5983ac08105ef213efb', 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'name': '/fluentd', 'state': 'running', 'status': 'Up 29 minutes'}) \u001b[0m",
                                "",
                                "TASK [Get count of ceph-osd containers on host] ********************************",
                                "Saturday 21 March 2026  01:17:18 +0000 (0:00:00.665)       0:00:04.841 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to failed when count of containers is wrong] *************",
                                "Saturday 21 March 2026  01:17:18 +0000 (0:00:00.316)       0:00:05.158 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if count matches] ******************************",
                                "Saturday 21 March 2026  01:17:18 +0000 (0:00:00.294)       0:00:05.452 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Saturday 21 March 2026  01:17:19 +0000 (0:00:00.333)       0:00:05.786 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get list of ceph-osd containers that are not running] ********************",
                                "Saturday 21 March 2026  01:17:19 +0000 (0:00:00.424)       0:00:06.210 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'ceph-osd-3', 'osd_id': '3', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'ceph-osd-0', 'osd_id': '0', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'ceph-osd-5', 'osd_id': '5', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'ceph-osd-1', 'osd_id': '1', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'ceph-osd-2', 'osd_id': '2', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'ceph-osd-4', 'osd_id': '4', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get count of ceph-osd containers that are not running] *******************",
                                "Saturday 21 March 2026  01:17:20 +0000 (0:00:00.322)       0:00:06.533 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if an OSD is not running] **********************",
                                "Saturday 21 March 2026  01:17:20 +0000 (0:00:00.284)       0:00:06.817 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if an OSD is not running] **********************",
                                "Saturday 21 March 2026  01:17:20 +0000 (0:00:00.293)       0:00:07.111 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if all containers are running] *****************",
                                "Saturday 21 March 2026  01:17:21 +0000 (0:00:00.480)       0:00:07.592 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Saturday 21 March 2026  01:17:21 +0000 (0:00:00.284)       0:00:07.876 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Saturday 21 March 2026  01:17:21 +0000 (0:00:00.241)       0:00:08.118 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Saturday 21 March 2026  01:17:21 +0000 (0:00:00.249)       0:00:08.368 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Saturday 21 March 2026  01:17:22 +0000 (0:00:00.239)       0:00:08.607 ******** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Saturday 21 March 2026  01:17:22 +0000 (0:00:00.068)       0:00:08.676 ******** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Saturday 21 March 2026  01:17:22 +0000 (0:00:00.067)       0:00:08.743 ******** ",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Saturday 21 March 2026  01:17:22 +0000 (0:00:00.068)       0:00:08.811 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail early due to containers not running] ********************************",
                                "Saturday 21 March 2026  01:17:22 +0000 (0:00:00.593)       0:00:09.405 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Saturday 21 March 2026  01:17:23 +0000 (0:00:00.247)       0:00:09.652 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set _mon_hostname fact] **************************************************",
                                "Saturday 21 March 2026  01:17:23 +0000 (0:00:00.305)       0:00:09.957 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Get ceph osd tree] *******************************************************",
                                "Saturday 21 March 2026  01:17:23 +0000 (0:00:00.228)       0:00:10.186 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [Parse osd tree from JSON] ************************************************",
                                "Saturday 21 March 2026  01:17:25 +0000 (0:00:02.012)       0:00:12.198 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Get OSDs that are not up or in] ******************************************",
                                "Saturday 21 March 2026  01:17:25 +0000 (0:00:00.116)       0:00:12.315 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail test if OSDs are not up or in] **************************************",
                                "Saturday 21 March 2026  01:17:26 +0000 (0:00:00.280)       0:00:12.595 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Pass test if OSDs are all up and in] *************************************",
                                "Saturday 21 March 2026  01:17:26 +0000 (0:00:00.113)       0:00:12.709 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Saturday 21 March 2026  01:17:26 +0000 (0:00:00.126)       0:00:12.836 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [List ceph LVM volumes and collect data] **********************************",
                                "Saturday 21 March 2026  01:17:26 +0000 (0:00:00.439)       0:00:13.276 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Parse LVM data as JSON] **************************************************",
                                "Saturday 21 March 2026  01:17:29 +0000 (0:00:02.505)       0:00:15.782 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get unencrypted and encrypted OSDs] **************************************",
                                "Saturday 21 March 2026  01:17:29 +0000 (0:00:00.322)       0:00:16.105 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if count of encrypted OSDs does not match] **************************",
                                "Saturday 21 March 2026  01:17:30 +0000 (0:00:00.450)       0:00:16.555 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Pass if count of encrypted OSDs equals count of OSDs] ********************",
                                "Saturday 21 March 2026  01:17:30 +0000 (0:00:00.473)       0:00:17.028 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if count of unencrypted OSDs does not match] ************************",
                                "Saturday 21 March 2026  01:17:30 +0000 (0:00:00.303)       0:00:17.331 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Pass if count of unencrypted OSDs equals count of OSDs] ******************",
                                "Saturday 21 March 2026  01:17:31 +0000 (0:00:00.283)       0:00:17.615 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Saturday 21 March 2026  01:17:31 +0000 (0:00:00.435)       0:00:18.051 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get CRUSH node data of each OSD host and root node childs] ***************",
                                "Saturday 21 March 2026  01:17:32 +0000 (0:00:00.485)       0:00:18.536 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate sub test expression results] ***********************************",
                                "Saturday 21 March 2026  01:17:32 +0000 (0:00:00.460)       0:00:18.997 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail test if any sub test failed] ****************************************",
                                "Saturday 21 March 2026  01:17:32 +0000 (0:00:00.294)       0:00:19.292 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Pass test if no sub test failed] *****************************************",
                                "Saturday 21 March 2026  01:17:33 +0000 (0:00:00.454)       0:00:19.746 ******** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set validation result to passed if no test failed] ***********************",
                                "Saturday 21 March 2026  01:17:33 +0000 (0:00:00.293)       0:00:20.040 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Set validation result to failed if a test failed] ************************",
                                "Saturday 21 March 2026  01:17:33 +0000 (0:00:00.249)       0:00:20.290 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Saturday 21 March 2026  01:17:34 +0000 (0:00:00.238)       0:00:20.529 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Saturday 21 March 2026  01:17:35 +0000 (0:00:01.716)       0:00:22.246 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Saturday 21 March 2026  01:17:36 +0000 (0:00:00.257)       0:00:22.503 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Saturday 21 March 2026  01:17:36 +0000 (0:00:00.250)       0:00:22.753 ******** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Saturday 21 March 2026  01:17:36 +0000 (0:00:00.068)       0:00:22.822 ******** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Saturday 21 March 2026  01:17:36 +0000 (0:00:00.216)       0:00:23.038 ******** ",
                                "",
                                "RUNNING HANDLER [Write report file] ********************************************",
                                "Saturday 21 March 2026  01:17:36 +0000 (0:00:00.070)       0:00:23.109 ******** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Saturday 21 March 2026  01:17:37 +0000 (0:00:01.206)       0:00:24.315 ******** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": [\u001b[0m",
                                "\u001b[0;32m        \"Validator run completed.\",\u001b[0m",
                                "\u001b[0;32m        \"You can find the report file here:\",\u001b[0m",
                                "\u001b[0;32m        \"/opt/reports/validator/ceph-osds-validator-2026-03-21T01:17:14+00:00-report.json\",\u001b[0m",
                                "\u001b[0;32m        \"on the following host:\",\u001b[0m",
                                "\u001b[0;32m        \"testbed-manager\"\u001b[0m",
                                "\u001b[0;32m    ]\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=35  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  01:17:38 +0000 (0:00:00.383)       0:00:24.698 ******** ",
                                "=============================================================================== ",
                                "List ceph LVM volumes and collect data ---------------------------------- 2.51s",
                                "Get ceph osd tree ------------------------------------------------------- 2.01s",
                                "Aggregate test results step one ----------------------------------------- 1.72s",
                                "Write report file ------------------------------------------------------- 1.21s",
                                "Get timestamp for report file ------------------------------------------- 0.95s",
                                "Create report output directory ------------------------------------------ 0.67s",
                                "Get list of ceph-osd containers on host --------------------------------- 0.67s",
                                "Print report file information ------------------------------------------- 0.59s",
                                "Prepare test data ------------------------------------------------------- 0.49s",
                                "Set test result to failed if an OSD is not running ---------------------- 0.48s",
                                "Fail if count of encrypted OSDs does not match -------------------------- 0.47s",
                                "Get CRUSH node data of each OSD host and root node childs --------------- 0.46s",
                                "Fail test if any sub test failed ---------------------------------------- 0.45s",
                                "Calculate OSD devices for each host ------------------------------------- 0.45s",
                                "Get unencrypted and encrypted OSDs -------------------------------------- 0.45s",
                                "Prepare test data ------------------------------------------------------- 0.44s",
                                "Pass if count of unencrypted OSDs equals count of OSDs ------------------ 0.44s",
                                "Prepare test data ------------------------------------------------------- 0.42s",
                                "Print report file information ------------------------------------------- 0.38s",
                                "Calculate OSD devices for each host ------------------------------------- 0.36s",
                                "+ sh -c /opt/configuration/scripts/check/200-infrastructure.sh",
                                "+ set -e",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=true",
                                "++ TEMPEST=true",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ [[ -e /etc/redhat-release ]]",
                                "+ source /etc/os-release",
                                "++ PRETTY_NAME='Ubuntu 24.04.4 LTS'",
                                "++ NAME=Ubuntu",
                                "++ VERSION_ID=24.04",
                                "++ VERSION='24.04.4 LTS (Noble Numbat)'",
                                "++ VERSION_CODENAME=noble",
                                "++ ID=ubuntu",
                                "++ ID_LIKE=debian",
                                "++ HOME_URL=https://www.ubuntu.com/",
                                "++ SUPPORT_URL=https://help.ubuntu.com/",
                                "++ BUG_REPORT_URL=https://bugs.launchpad.net/ubuntu/",
                                "++ PRIVACY_POLICY_URL=https://www.ubuntu.com/legal/terms-and-policies/privacy-policy",
                                "++ UBUNTU_CODENAME=noble",
                                "++ LOGO=ubuntu-logo",
                                "+ [[ ubuntu == \\u\\b\\u\\n\\t\\u ]]",
                                "+ packages='libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client'",
                                "+ dpkg -s libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client",
                                "+ sudo apt-get install -y libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client",
                                "",
                                "# Status of Elasticsearch",
                                "",
                                "+ pushd /opt/configuration/contrib",
                                "+ echo",
                                "+ echo '# Status of Elasticsearch'",
                                "+ echo",
                                "+ bash nagios-plugins/check_elasticsearch -H api-int.testbed.osism.xyz -s",
                                "OK - elasticsearch (kolla_logging) is running. status: green; timed_out: false; number_of_nodes: 3; number_of_data_nodes: 3; active_primary_shards: 9; active_shards: 22; relocating_shards: 0; initializing_shards: 0; delayed_unassigned_shards: 0; unassigned_shards: 0  | 'active_primary'=9 'active'=22 'relocating'=0 'init'=0 'delay_unass'=0 'unass'=0",
                                "",
                                "# Status of MariaDB",
                                "",
                                "+ echo",
                                "+ echo '# Status of MariaDB'",
                                "+ echo",
                                "++ semver latest 10.0.0-0",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ osism status database",
                                "\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConnecting to MariaDB at 192.168.16.9 as root_shard_0...\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster Status: Primary\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConnected: ON\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mReady: ON\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster Size: 3\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLocal State: Synced\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster State UUID: 7e30386f-24c0-11f1-a9de-baebd475a8d6\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster Members: 192.168.16.11:3306,192.168.16.12:3306,192.168.16.10:3306\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGalera Version: 26.4.25(r7387a566)\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLocal Node UUID: b04005f7-24c0-11f1-9f1a-2e2443d031b0\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlow Control Paused: 0.00%\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRecv Queue Avg: 0.00884956\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSend Queue Avg: 0.000591366\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTransactions: 4520 local commits, 6705 replicated, 113 received\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConflicts: 0 cert failures, 0 bf aborts\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMariaDB Uptime: 22 minutes, 21 seconds\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mThreads: 126 connected, 1 running\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mQueries: 237598 total, 0 slow\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAborted Connects: 120\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMariaDB Galera Cluster validation PASSED\u001b[0m",
                                "",
                                "# Status of Prometheus",
                                "",
                                "+ echo",
                                "+ echo '# Status of Prometheus'",
                                "+ echo",
                                "+ curl -s https://api-int.testbed.osism.xyz:9091/-/healthy",
                                "Unauthorized",
                                "+ curl -s https://api-int.testbed.osism.xyz:9091/-/ready",
                                "Unauthorized",
                                "",
                                "# Status of RabbitMQ",
                                "",
                                "+ echo",
                                "+ echo '# Status of RabbitMQ'",
                                "+ echo",
                                "++ semver latest 10.0.0-0",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ osism status messaging",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Connecting to RabbitMQ Management API at 192.168.16.10:15672 as openstack...\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] RabbitMQ Version: 3.13.7\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Erlang Version: 26.2.5.15\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Cluster Name: rabbit@testbed-node-0\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Cluster Size: 3\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Running Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Partitions: None (healthy)\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Connections: 205, Channels: 204, Queues: 173\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Messages: 218 total, 218 ready, 0 unacked\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Message Rates: 7.8/s publish, 8.2/s deliver\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Disk Free: 58.1 GB (limit: 0.0 GB)\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Memory Used: 0.18 GB (limit: 12.54 GB)\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] File Descriptors: 116/1024\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Sockets: 68/832\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Connecting to RabbitMQ Management API at 192.168.16.11:15672 as openstack...\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] RabbitMQ Version: 3.13.7\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Erlang Version: 26.2.5.15\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Cluster Name: rabbit@testbed-node-1\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Cluster Size: 3\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Running Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Partitions: None (healthy)\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Connections: 205, Channels: 204, Queues: 173\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Messages: 218 total, 218 ready, 0 unacked\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Message Rates: 7.8/s publish, 8.2/s deliver\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Disk Free: 58.4 GB (limit: 0.0 GB)\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Memory Used: 0.18 GB (limit: 12.54 GB)\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] File Descriptors: 124/1024\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Sockets: 78/832\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Connecting to RabbitMQ Management API at 192.168.16.12:15672 as openstack...\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] RabbitMQ Version: 3.13.7\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Erlang Version: 26.2.5.15\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Cluster Name: rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Cluster Size: 3\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Running Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Partitions: None (healthy)\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Connections: 205, Channels: 204, Queues: 173\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Messages: 218 total, 218 ready, 0 unacked\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Message Rates: 7.8/s publish, 8.2/s deliver\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Disk Free: 58.4 GB (limit: 0.0 GB)\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Memory Used: 0.17 GB (limit: 12.54 GB)\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] File Descriptors: 105/1024\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Sockets: 59/832\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRabbitMQ Cluster validation PASSED\u001b[0m",
                                "",
                                "# Status of Redis",
                                "",
                                "+ echo",
                                "+ echo '# Status of Redis'",
                                "+ echo",
                                "+ /usr/lib/nagios/plugins/check_tcp -H 192.168.16.10 -p 6379 -A -E -s 'AUTH QHNA1SZRlOKzLADhUd5ZDgpHfQe6dNfr3bwEdY24\\r\\nPING\\r\\nINFO replication\\r\\nQUIT\\r\\n' -e PONG -e role:master -e slave0:ip=192.168.16.1 -e,port=6379 -j",
                                "TCP OK - 0.001 second response time on 192.168.16.10 port 6379|time=0.001083s;;;0.000000;10.000000",
                                "",
                                "+ popd",
                                "+ echo",
                                "# Create backup of MariaDB database",
                                "+ echo '# Create backup of MariaDB database'",
                                "+ echo",
                                "",
                                "+ osism apply mariadb_backup -e mariadb_backup_type=full",
                                "\u001b[32m2026-03-21 01:18:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of mariadb_backup.\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09f29ddb-90dc-4859-b8a6-c08db303db76 (mariadb_backup) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 01:18:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 09f29ddb-90dc-4859-b8a6-c08db303db76 (mariadb_backup) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Saturday 21 March 2026  01:18:16 +0000 (0:00:00.237)       0:00:00.237 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Saturday 21 March 2026  01:18:16 +0000 (0:00:00.299)       0:00:00.536 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_mariadb_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_mariadb_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_mariadb_True)\u001b[0m",
                                "",
                                "PLAY [Apply role mariadb] ******************************************************",
                                "",
                                "TASK [mariadb : Group MariaDB hosts based on shards] ***************************",
                                "Saturday 21 March 2026  01:18:17 +0000 (0:00:00.407)       0:00:00.944 ******** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Saturday 21 March 2026  01:18:17 +0000 (0:00:00.402)       0:00:01.347 ******** ",
                                "\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/backup.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [mariadb : Get MariaDB container facts] ***********************************",
                                "Saturday 21 March 2026  01:18:18 +0000 (0:00:00.623)       0:00:01.971 ******** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Taking full database backup via Mariabackup] *******************",
                                "Saturday 21 March 2026  01:18:21 +0000 (0:00:03.368)       0:00:05.340 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_restart\u001b[0m",
                                "",
                                "PLAY [Restart mariadb services] ************************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_start\u001b[0m",
                                "",
                                "PLAY [Start mariadb services] **************************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m",
                                "\u001b[1;35mmariadb_bootstrap_restart\u001b[0m",
                                "",
                                "PLAY [Restart bootstrap mariadb service] ***************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Apply mariadb post-configuration] ****************************************",
                                "",
                                "TASK [Include mariadb post-deploy.yml] *****************************************",
                                "Saturday 21 March 2026  01:20:00 +0000 (0:01:39.326)       0:01:44.667 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Include mariadb post-upgrade.yml] ****************************************",
                                "Saturday 21 March 2026  01:20:01 +0000 (0:00:00.277)       0:01:44.944 ******** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=6   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=4   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=4   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  01:20:01 +0000 (0:00:00.212)       0:01:45.157 ******** ",
                                "=============================================================================== ",
                                "mariadb : Taking full database backup via Mariabackup ------------------ 99.33s",
                                "mariadb : Get MariaDB container facts ----------------------------------- 3.37s",
                                "mariadb : include_tasks ------------------------------------------------- 0.62s",
                                "Group hosts based on enabled services ----------------------------------- 0.41s",
                                "mariadb : Group MariaDB hosts based on shards --------------------------- 0.40s",
                                "Group hosts based on Kolla action --------------------------------------- 0.30s",
                                "Include mariadb post-deploy.yml ----------------------------------------- 0.28s",
                                "Include mariadb post-upgrade.yml ---------------------------------------- 0.21s",
                                "+ sh -c /opt/configuration/scripts/check/300-openstack.sh",
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/configuration/scripts/manager-version.sh",
                                "+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml",
                                "",
                                "# OpenStack endpoints",
                                "",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "+ export OS_CLOUD=admin",
                                "+ OS_CLOUD=admin",
                                "+ echo",
                                "+ echo '# OpenStack endpoints'",
                                "+ echo",
                                "+ openstack endpoint list",
                                "+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+",
                                "| ID                               | Region    | Service Name | Service Type    | Enabled | Interface | URL                                                                 |",
                                "+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+",
                                "| 045abbc9413e49eab3d35b5b46ac273d | RegionOne | octavia      | load-balancer   | True    | internal  | https://api-int.testbed.osism.xyz:9876                              |",
                                "| 1d06f6cc268149419c413b9ef1f0325f | RegionOne | barbican     | key-manager     | True    | internal  | https://api-int.testbed.osism.xyz:9311                              |",
                                "| 205ecdb1bc334998aac9bdfbb6df373a | RegionOne | nova         | compute         | True    | internal  | https://api-int.testbed.osism.xyz:8774/v2.1                         |",
                                "| 48009575ba874f9881df4ddc8a9306f9 | RegionOne | swift        | object-store    | True    | internal  | https://api-int.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s |",
                                "| 48e1442c694a4a8fa468b7430f3c0f24 | RegionOne | barbican     | key-manager     | True    | public    | https://api.testbed.osism.xyz:9311                                  |",
                                "| 577becedbb42491e82fb5b2dea82a3ee | RegionOne | glance       | image           | True    | public    | https://api.testbed.osism.xyz:9292                                  |",
                                "| 57daba2f887740ba8be672a34321ba57 | RegionOne | placement    | placement       | True    | public    | https://api.testbed.osism.xyz:8780                                  |",
                                "| 59fb3fb105524391bae1670c0e48d706 | RegionOne | magnum       | container-infra | True    | internal  | https://api-int.testbed.osism.xyz:9511/v1                           |",
                                "| 6ba274fecd534959bf94f16b1484f3ed | RegionOne | cinderv3     | volumev3        | True    | public    | https://api.testbed.osism.xyz:8776/v3/%(tenant_id)s                 |",
                                "| 84ff3c7b93ca4a4799e5406c09d29dbf | RegionOne | placement    | placement       | True    | internal  | https://api-int.testbed.osism.xyz:8780                              |",
                                "| 86304183b84b4ebdbca0cb27fc1aa03c | RegionOne | magnum       | container-infra | True    | public    | https://api.testbed.osism.xyz:9511/v1                               |",
                                "| b060be793e0c48a5a67a8a574e893250 | RegionOne | designate    | dns             | True    | public    | https://api.testbed.osism.xyz:9001                                  |",
                                "| bb106454a2a9409f9bd8ce1b5864e0ef | RegionOne | designate    | dns             | True    | internal  | https://api-int.testbed.osism.xyz:9001                              |",
                                "| bece2b03d3824cd1b7f3e45a5a595172 | RegionOne | glance       | image           | True    | internal  | https://api-int.testbed.osism.xyz:9292                              |",
                                "| bf1616dcfb4647999288d99d32b6f9c6 | RegionOne | keystone     | identity        | True    | internal  | https://api-int.testbed.osism.xyz:5000                              |",
                                "| db9674b056614c71a5ed51993fbb258e | RegionOne | neutron      | network         | True    | internal  | https://api-int.testbed.osism.xyz:9696                              |",
                                "| e0e8429d47034e81be4c454687ce067b | RegionOne | swift        | object-store    | True    | public    | https://api.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s     |",
                                "| e0f40e46c018498d99911bcbc5191bf3 | RegionOne | keystone     | identity        | True    | public    | https://api.testbed.osism.xyz:5000                                  |",
                                "| e3a16ff8f02c4c3faa07372cf5be5f4e | RegionOne | octavia      | load-balancer   | True    | public    | https://api.testbed.osism.xyz:9876                                  |",
                                "| e6534e6bf126456da312787bee2f2088 | RegionOne | nova         | compute         | True    | public    | https://api.testbed.osism.xyz:8774/v2.1                             |",
                                "| eba1df94bb6d45638c30071431b6fdee | RegionOne | cinderv3     | volumev3        | True    | internal  | https://api-int.testbed.osism.xyz:8776/v3/%(tenant_id)s             |",
                                "| ece1fe5797a742e09c884f898f3f4451 | RegionOne | neutron      | network         | True    | public    | https://api.testbed.osism.xyz:9696                                  |",
                                "+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+",
                                "",
                                "# Cinder",
                                "",
                                "+ echo",
                                "+ echo '# Cinder'",
                                "+ echo",
                                "+ openstack volume service list",
                                "+------------------+----------------------------+----------+---------+-------+----------------------------+",
                                "| Binary           | Host                       | Zone     | Status  | State | Updated At                 |",
                                "+------------------+----------------------------+----------+---------+-------+----------------------------+",
                                "| cinder-scheduler | testbed-node-0             | internal | enabled | up    | 2026-03-21T01:20:02.000000 |",
                                "| cinder-scheduler | testbed-node-2             | internal | enabled | up    | 2026-03-21T01:20:01.000000 |",
                                "| cinder-scheduler | testbed-node-1             | internal | enabled | up    | 2026-03-21T01:20:02.000000 |",
                                "| cinder-volume    | testbed-node-0@rbd-volumes | nova     | enabled | up    | 2026-03-21T01:20:01.000000 |",
                                "| cinder-volume    | testbed-node-1@rbd-volumes | nova     | enabled | up    | 2026-03-21T01:20:03.000000 |",
                                "| cinder-volume    | testbed-node-2@rbd-volumes | nova     | enabled | up    | 2026-03-21T01:20:04.000000 |",
                                "| cinder-backup    | testbed-node-0             | nova     | enabled | up    | 2026-03-21T01:20:07.000000 |",
                                "| cinder-backup    | testbed-node-2             | nova     | enabled | up    | 2026-03-21T01:19:58.000000 |",
                                "| cinder-backup    | testbed-node-1             | nova     | enabled | up    | 2026-03-21T01:19:58.000000 |",
                                "+------------------+----------------------------+----------+---------+-------+----------------------------+",
                                "",
                                "# Neutron",
                                "",
                                "+ echo",
                                "+ echo '# Neutron'",
                                "+ echo",
                                "+ openstack network agent list",
                                "+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+",
                                "| ID                                   | Agent Type                   | Host           | Availability Zone | Alive | State | Binary                     |",
                                "+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+",
                                "| testbed-node-2                       | OVN Controller Gateway agent | testbed-node-2 | nova              | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-5                       | OVN Controller agent         | testbed-node-5 |                   | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-3                       | OVN Controller agent         | testbed-node-3 |                   | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-0                       | OVN Controller Gateway agent | testbed-node-0 | nova              | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-4                       | OVN Controller agent         | testbed-node-4 |                   | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-1                       | OVN Controller Gateway agent | testbed-node-1 | nova              | :-)   | UP    | ovn-controller             |",
                                "| e645415a-98f5-5758-8cd1-c47af282b5c0 | OVN Metadata agent           | testbed-node-3 |                   | :-)   | UP    | neutron-ovn-metadata-agent |",
                                "| 36b9d21c-9928-5c0a-9b27-73ac7a3e770c | OVN Metadata agent           | testbed-node-5 |                   | :-)   | UP    | neutron-ovn-metadata-agent |",
                                "| 4939696e-6092-5a33-bb73-b850064684df | OVN Metadata agent           | testbed-node-4 |                   | :-)   | UP    | neutron-ovn-metadata-agent |",
                                "+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+",
                                "+ openstack network service provider list",
                                "+---------------+------+---------+",
                                "| Service Type  | Name | Default |",
                                "+---------------+------+---------+",
                                "| L3_ROUTER_NAT | ovn  | True    |",
                                "+---------------+------+---------+",
                                "",
                                "+ echo",
                                "+ echo '# Nova'",
                                "# Nova",
                                "+ echo",
                                "",
                                "+ openstack compute service list",
                                "+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+",
                                "| ID                                   | Binary         | Host           | Zone     | Status  | State | Updated At                 |",
                                "+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+",
                                "| 2dfd5f0f-b0d5-4e16-b497-90455e00daac | nova-scheduler | testbed-node-0 | internal | enabled | up    | 2026-03-21T01:20:15.000000 |",
                                "| 66935adb-cd12-41ae-9556-8fe2ef82f37d | nova-scheduler | testbed-node-2 | internal | enabled | up    | 2026-03-21T01:20:16.000000 |",
                                "| bbc1bc14-a891-4a3b-95ca-4222f02fc912 | nova-scheduler | testbed-node-1 | internal | enabled | up    | 2026-03-21T01:20:06.000000 |",
                                "| 62c71ef9-898e-4407-85e2-b6a57bf4b607 | nova-conductor | testbed-node-0 | internal | enabled | up    | 2026-03-21T01:20:08.000000 |",
                                "| ecb7670e-6157-4994-b001-5d6e1cd0b540 | nova-conductor | testbed-node-2 | internal | enabled | up    | 2026-03-21T01:20:09.000000 |",
                                "| 40f2b6fa-67a1-4318-8e28-dec815488cb1 | nova-conductor | testbed-node-1 | internal | enabled | up    | 2026-03-21T01:20:09.000000 |",
                                "| 31494761-ad41-4745-9b1c-34b1a54d475e | nova-compute   | testbed-node-3 | nova     | enabled | up    | 2026-03-21T01:20:07.000000 |",
                                "| 692abb40-5d07-4597-8392-29ef6c60842a | nova-compute   | testbed-node-4 | nova     | enabled | up    | 2026-03-21T01:20:08.000000 |",
                                "| 06ef3854-d0a7-429b-9792-9aa3169e2ca2 | nova-compute   | testbed-node-5 | nova     | enabled | up    | 2026-03-21T01:20:07.000000 |",
                                "+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+",
                                "+ openstack hypervisor list",
                                "+--------------------------------------+---------------------+-----------------+---------------+-------+",
                                "| ID                                   | Hypervisor Hostname | Hypervisor Type | Host IP       | State |",
                                "+--------------------------------------+---------------------+-----------------+---------------+-------+",
                                "| 49c11798-5996-4b3e-a240-8a1cf8c1b2f5 | testbed-node-3      | QEMU            | 192.168.16.13 | up    |",
                                "| 856225ce-2980-4f96-85fa-cef245de4a09 | testbed-node-5      | QEMU            | 192.168.16.15 | up    |",
                                "| b98edce9-70d1-426b-a3d6-43717913d625 | testbed-node-4      | QEMU            | 192.168.16.14 | up    |",
                                "+--------------------------------------+---------------------+-----------------+---------------+-------+",
                                "",
                                "# Run OpenStack test play",
                                "",
                                "+ echo",
                                "+ echo '# Run OpenStack test play'",
                                "+ echo",
                                "+ osism apply --environment openstack test",
                                "\u001b[32m2026-03-21 01:20:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play test in environment openstack\u001b[0m",
                                "\u001b[32m2026-03-21 01:20:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of test.\u001b[0m",
                                "\u001b[32m2026-03-21 01:20:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0c177e24-c00d-4262-a7a1-a17b71dc28f9 (test) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 01:20:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 0c177e24-c00d-4262-a7a1-a17b71dc28f9 (test) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Create test project] *****************************************************",
                                "",
                                "TASK [Create test domain] ******************************************************",
                                "Saturday 21 March 2026  01:20:25 +0000 (0:00:00.103)       0:00:00.103 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test-admin user] **************************************************",
                                "Saturday 21 March 2026  01:20:28 +0000 (0:00:03.558)       0:00:03.662 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Add manager role to user test-admin] *************************************",
                                "Saturday 21 March 2026  01:20:32 +0000 (0:00:03.828)       0:00:07.491 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test project] *****************************************************",
                                "Saturday 21 March 2026  01:20:38 +0000 (0:00:06.134)       0:00:13.625 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test user] ********************************************************",
                                "Saturday 21 March 2026  01:20:42 +0000 (0:00:04.213)       0:00:17.838 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Add member roles to user test] *******************************************",
                                "Saturday 21 March 2026  01:20:46 +0000 (0:00:04.111)       0:00:21.950 ******** ",
                                "\u001b[0;33mchanged: [localhost] => (item=load-balancer_member)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=member)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=creator)\u001b[0m",
                                "",
                                "TASK [Create test server group] ************************************************",
                                "Saturday 21 March 2026  01:20:58 +0000 (0:00:11.821)       0:00:33.771 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create ssh security group] ***********************************************",
                                "Saturday 21 March 2026  01:21:03 +0000 (0:00:04.407)       0:00:38.178 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Add rule to ssh security group] ******************************************",
                                "Saturday 21 March 2026  01:21:07 +0000 (0:00:04.579)       0:00:42.758 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create icmp security group] **********************************************",
                                "Saturday 21 March 2026  01:21:11 +0000 (0:00:04.120)       0:00:46.878 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Add rule to icmp security group] *****************************************",
                                "Saturday 21 March 2026  01:21:15 +0000 (0:00:04.101)       0:00:50.980 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test keypair] *****************************************************",
                                "Saturday 21 March 2026  01:21:20 +0000 (0:00:04.105)       0:00:55.085 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test network] *****************************************************",
                                "Saturday 21 March 2026  01:21:23 +0000 (0:00:03.968)       0:00:59.054 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test subnet] ******************************************************",
                                "Saturday 21 March 2026  01:21:29 +0000 (0:00:05.281)       0:01:04.335 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test router] ******************************************************",
                                "Saturday 21 March 2026  01:21:35 +0000 (0:00:05.971)       0:01:10.307 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "PLAY [Manage test instances and volumes] ***************************************",
                                "",
                                "TASK [Get test server group] ***************************************************",
                                "Saturday 21 March 2026  01:21:45 +0000 (0:00:10.595)       0:01:20.902 ******** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "TASK [Detach test volume] ******************************************************",
                                "Saturday 21 March 2026  01:21:49 +0000 (0:00:03.387)       0:01:24.289 ******** ",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "TASK [Delete test volume] ******************************************************",
                                "Saturday 21 March 2026  01:21:49 +0000 (0:00:00.057)       0:01:24.347 ******** ",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "TASK [Delete test instances] ***************************************************",
                                "Saturday 21 March 2026  01:21:49 +0000 (0:00:00.075)       0:01:24.422 ******** ",
                                "\u001b[0;36mskipping: [localhost] => (item=test-4) \u001b[0m",
                                "\u001b[0;36mskipping: [localhost] => (item=test-3) \u001b[0m",
                                "\u001b[0;36mskipping: [localhost] => (item=test-2) \u001b[0m",
                                "\u001b[0;36mskipping: [localhost] => (item=test-1) \u001b[0m",
                                "\u001b[0;36mskipping: [localhost] => (item=test) \u001b[0m",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "TASK [Wait for instance deletion to complete] **********************************",
                                "Saturday 21 March 2026  01:21:49 +0000 (0:00:00.143)       0:01:24.565 ******** ",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test instances] ***************************************************",
                                "Saturday 21 March 2026  01:21:49 +0000 (0:00:00.124)       0:01:24.690 ******** ",
                                "\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m",
                                "",
                                "TASK [Wait for instance creation to complete] **********************************",
                                "Saturday 21 March 2026  01:21:53 +0000 (0:00:04.086)       0:01:28.776 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (60 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (59 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (58 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (57 retries left).\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j114091529227.2736', 'results_file': '/ansible/.ansible_async/j114091529227.2736', 'changed': True, 'item': 'test', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j550367883200.2761', 'results_file': '/ansible/.ansible_async/j550367883200.2761', 'changed': True, 'item': 'test-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j292689331656.2786', 'results_file': '/ansible/.ansible_async/j292689331656.2786', 'changed': True, 'item': 'test-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j923253320887.2818', 'results_file': '/ansible/.ansible_async/j923253320887.2818', 'changed': True, 'item': 'test-3', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j335587441571.2843', 'results_file': '/ansible/.ansible_async/j335587441571.2843', 'changed': True, 'item': 'test-4', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [Add metadata to instances] ***********************************************",
                                "Saturday 21 March 2026  01:22:40 +0000 (0:00:47.008)       0:02:15.785 ******** ",
                                "\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m",
                                "",
                                "TASK [Wait for metadata to be added] *******************************************",
                                "Saturday 21 March 2026  01:22:45 +0000 (0:00:04.580)       0:02:20.365 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for metadata to be added (30 retries left).\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j529373546002.2940', 'results_file': '/ansible/.ansible_async/j529373546002.2940', 'changed': True, 'item': 'test', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j977543273719.2965', 'results_file': '/ansible/.ansible_async/j977543273719.2965', 'changed': True, 'item': 'test-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j868334366829.2990', 'results_file': '/ansible/.ansible_async/j868334366829.2990', 'changed': True, 'item': 'test-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j116173683184.3015', 'results_file': '/ansible/.ansible_async/j116173683184.3015', 'changed': True, 'item': 'test-3', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j213789100609.3040', 'results_file': '/ansible/.ansible_async/j213789100609.3040', 'changed': True, 'item': 'test-4', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [Add tag to instances] ****************************************************",
                                "Saturday 21 March 2026  01:22:54 +0000 (0:00:09.169)       0:02:29.535 ******** ",
                                "\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m",
                                "",
                                "TASK [Wait for tags to be added] ***********************************************",
                                "Saturday 21 March 2026  01:22:58 +0000 (0:00:04.306)       0:02:33.842 ******** ",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for tags to be added (30 retries left).\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j164432034846.3116', 'results_file': '/ansible/.ansible_async/j164432034846.3116', 'changed': True, 'item': 'test', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j544760755925.3141', 'results_file': '/ansible/.ansible_async/j544760755925.3141', 'changed': True, 'item': 'test-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j529594319159.3167', 'results_file': '/ansible/.ansible_async/j529594319159.3167', 'changed': True, 'item': 'test-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j894484329495.3193', 'results_file': '/ansible/.ansible_async/j894484329495.3193', 'changed': True, 'item': 'test-3', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j226104707254.3219', 'results_file': '/ansible/.ansible_async/j226104707254.3219', 'changed': True, 'item': 'test-4', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [Create test volume] ******************************************************",
                                "Saturday 21 March 2026  01:23:08 +0000 (0:00:09.356)       0:02:43.199 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Attach test volume] ******************************************************",
                                "Saturday 21 March 2026  01:23:14 +0000 (0:00:06.545)       0:02:49.745 ******** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create floating ip address] **********************************************",
                                "Saturday 21 March 2026  01:23:28 +0000 (0:00:13.448)       0:03:03.193 ******** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "TASK [Print floating ip address] ***********************************************",
                                "Saturday 21 March 2026  01:23:32 +0000 (0:00:04.489)       0:03:07.683 ******** ",
                                "\u001b[0;32mok: [localhost] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"192.168.112.150\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  01:23:32 +0000 (0:00:00.040)       0:03:07.724 ******** ",
                                "=============================================================================== ",
                                "Wait for instance creation to complete --------------------------------- 47.01s",
                                "Attach test volume ----------------------------------------------------- 13.45s",
                                "Add member roles to user test ------------------------------------------ 11.82s",
                                "Create test router ----------------------------------------------------- 10.60s",
                                "Wait for tags to be added ----------------------------------------------- 9.36s",
                                "Wait for metadata to be added ------------------------------------------- 9.17s",
                                "Create test volume ------------------------------------------------------ 6.55s",
                                "Add manager role to user test-admin ------------------------------------- 6.13s",
                                "Create test subnet ------------------------------------------------------ 5.97s",
                                "Create test network ----------------------------------------------------- 5.28s",
                                "Add metadata to instances ----------------------------------------------- 4.58s",
                                "Create ssh security group ----------------------------------------------- 4.58s",
                                "Create floating ip address ---------------------------------------------- 4.49s",
                                "Create test server group ------------------------------------------------ 4.41s",
                                "Add tag to instances ---------------------------------------------------- 4.31s",
                                "Create test project ----------------------------------------------------- 4.21s",
                                "Add rule to ssh security group ------------------------------------------ 4.12s",
                                "Create test user -------------------------------------------------------- 4.11s",
                                "Add rule to icmp security group ----------------------------------------- 4.11s",
                                "Create icmp security group ---------------------------------------------- 4.10s",
                                "+ server_list",
                                "+ openstack --os-cloud test server list",
                                "+--------------------------------------+--------+--------+---------------------------------------+--------------------------+----------+",
                                "| ID                                   | Name   | Status | Networks                              | Image                    | Flavor   |",
                                "+--------------------------------------+--------+--------+---------------------------------------+--------------------------+----------+",
                                "| e341f1ff-bd52-468b-8137-05723da11652 | test-4 | ACTIVE | test=192.168.112.145, 192.168.200.240 | N/A (booted from volume) | SCS-1L-1 |",
                                "| a219ecb7-4093-44bb-b939-a453113dfec2 | test-3 | ACTIVE | test=192.168.112.103, 192.168.200.23  | N/A (booted from volume) | SCS-1L-1 |",
                                "| ef7382d2-8009-467e-b354-44b46a384980 | test-2 | ACTIVE | test=192.168.112.117, 192.168.200.243 | N/A (booted from volume) | SCS-1L-1 |",
                                "| 4740b0e6-2766-4b54-a663-63e0e558e778 | test   | ACTIVE | test=192.168.112.150, 192.168.200.203 | N/A (booted from volume) | SCS-1L-1 |",
                                "| 56e68f34-0e54-4c15-a73d-f764465f4df9 | test-1 | ACTIVE | test=192.168.112.200, 192.168.200.144 | N/A (booted from volume) | SCS-1L-1 |",
                                "+--------------------------------------+--------+--------+---------------------------------------+--------------------------+----------+",
                                "+ openstack --os-cloud test server show test",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-03-21T01:22:24.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.150, 192.168.200.203                                                                                                                                                                                                                                                                                                                                               |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-03-21T01:21:58Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='True', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | 7db2ef38c4e8041c1ab68e4105682d7049680cf0587156722412a340                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | 4740b0e6-2766-4b54-a663-63e0e558e778                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | ce936a29388e4c2e9c53a72b276e2fb2                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test'                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| security_groups                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "|                                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-03-21T01:22:46Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | 413e209667b94f5e9151b49aab438337                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='383f89d6-5820-4fa6-9c5d-f29086bb0799'                                                                                                                                                                                                                                                                                                             |",
                                "|                                     | delete_on_termination='False', id='96ecdebe-a1b7-4c81-99fd-d6c165f8a05d'                                                                                                                                                                                                                                                                                                            |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ openstack --os-cloud test server show test-1",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test-1                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-03-21T01:22:24.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.200, 192.168.200.144                                                                                                                                                                                                                                                                                                                                               |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-03-21T01:21:58Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='True', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | 7db2ef38c4e8041c1ab68e4105682d7049680cf0587156722412a340                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | 56e68f34-0e54-4c15-a73d-f764465f4df9                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test-1                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | ce936a29388e4c2e9c53a72b276e2fb2                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test-1'                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| security_groups                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "|                                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-03-21T01:22:46Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | 413e209667b94f5e9151b49aab438337                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='bedb1497-49f1-4ba2-8d84-4c80dfa714cf'                                                                                                                                                                                                                                                                                                             |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ openstack --os-cloud test server show test-2",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test-2                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-03-21T01:22:26.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.117, 192.168.200.243                                                                                                                                                                                                                                                                                                                                               |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-03-21T01:21:59Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='True', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | 6ede240e48d06e40d4d6e3ffd457e1e0ed1e43d51a67d24c33c067cf                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | ef7382d2-8009-467e-b354-44b46a384980                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test-2                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | ce936a29388e4c2e9c53a72b276e2fb2                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test-2'                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| security_groups                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "|                                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-03-21T01:22:47Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | 413e209667b94f5e9151b49aab438337                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='39b5d04c-efff-4103-8694-5044b75d8ed3'                                                                                                                                                                                                                                                                                                             |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ openstack --os-cloud test server show test-3",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test-3                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-03-21T01:22:26.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.103, 192.168.200.23                                                                                                                                                                                                                                                                                                                                                |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-03-21T01:21:59Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='True', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | 6ede240e48d06e40d4d6e3ffd457e1e0ed1e43d51a67d24c33c067cf                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | a219ecb7-4093-44bb-b939-a453113dfec2                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test-3                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | ce936a29388e4c2e9c53a72b276e2fb2                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test-3'                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| security_groups                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "|                                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-03-21T01:22:48Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | 413e209667b94f5e9151b49aab438337                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='bf295bac-2ea3-4452-914f-011aaf87a95b'                                                                                                                                                                                                                                                                                                             |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ openstack --os-cloud test server show test-4",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test-4                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-03-21T01:22:25.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.145, 192.168.200.240                                                                                                                                                                                                                                                                                                                                               |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-03-21T01:22:01Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='True', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | 6ede240e48d06e40d4d6e3ffd457e1e0ed1e43d51a67d24c33c067cf                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | e341f1ff-bd52-468b-8137-05723da11652                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test-4                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | ce936a29388e4c2e9c53a72b276e2fb2                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test-4'                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| security_groups                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "|                                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-03-21T01:22:49Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | 413e209667b94f5e9151b49aab438337                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='c3f0bedd-88a5-41cd-992f-4319ff75bf1e'                                                                                                                                                                                                                                                                                                             |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ server_ping",
                                "++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'",
                                "++ tr -d '\\r'",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.103",
                                "PING 192.168.112.103 (192.168.112.103) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.103: icmp_seq=1 ttl=63 time=3.68 ms",
                                "64 bytes from 192.168.112.103: icmp_seq=2 ttl=63 time=1.68 ms",
                                "64 bytes from 192.168.112.103: icmp_seq=3 ttl=63 time=1.24 ms",
                                "",
                                "--- 192.168.112.103 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2004ms",
                                "rtt min/avg/max/mdev = 1.240/2.202/3.684/1.063 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.117",
                                "PING 192.168.112.117 (192.168.112.117) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.117: icmp_seq=1 ttl=63 time=4.32 ms",
                                "64 bytes from 192.168.112.117: icmp_seq=2 ttl=63 time=1.38 ms",
                                "64 bytes from 192.168.112.117: icmp_seq=3 ttl=63 time=1.09 ms",
                                "",
                                "--- 192.168.112.117 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2002ms",
                                "rtt min/avg/max/mdev = 1.091/2.262/4.318/1.458 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.150",
                                "PING 192.168.112.150 (192.168.112.150) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.150: icmp_seq=1 ttl=63 time=3.32 ms",
                                "64 bytes from 192.168.112.150: icmp_seq=2 ttl=63 time=1.29 ms",
                                "64 bytes from 192.168.112.150: icmp_seq=3 ttl=63 time=2.79 ms",
                                "",
                                "--- 192.168.112.150 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2002ms",
                                "rtt min/avg/max/mdev = 1.289/2.467/3.322/0.860 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.145",
                                "PING 192.168.112.145 (192.168.112.145) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.145: icmp_seq=1 ttl=63 time=3.60 ms",
                                "64 bytes from 192.168.112.145: icmp_seq=2 ttl=63 time=1.63 ms",
                                "64 bytes from 192.168.112.145: icmp_seq=3 ttl=63 time=0.967 ms",
                                "",
                                "--- 192.168.112.145 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 0.967/2.063/3.595/1.116 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.200",
                                "PING 192.168.112.200 (192.168.112.200) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.200: icmp_seq=1 ttl=63 time=4.57 ms",
                                "64 bytes from 192.168.112.200: icmp_seq=2 ttl=63 time=1.39 ms",
                                "64 bytes from 192.168.112.200: icmp_seq=3 ttl=63 time=1.33 ms",
                                "",
                                "--- 192.168.112.200 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.325/2.427/4.566/1.512 ms",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ compute_list",
                                "+ osism manage compute list testbed-node-3",
                                "\u001b[32m2026-03-21 01:24:05\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-21 01:24:05\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-21 01:24:05\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ osism manage compute list testbed-node-4",
                                "\u001b[32m2026-03-21 01:24:09\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-21 01:24:09\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-21 01:24:09\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| 4740b0e6-2766-4b54-a663-63e0e558e778 | test   | ACTIVE   |",
                                "| 56e68f34-0e54-4c15-a73d-f764465f4df9 | test-1 | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ osism manage compute list testbed-node-5",
                                "\u001b[32m2026-03-21 01:24:13\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-21 01:24:13\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-21 01:24:13\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| e341f1ff-bd52-468b-8137-05723da11652 | test-4 | ACTIVE   |",
                                "| a219ecb7-4093-44bb-b939-a453113dfec2 | test-3 | ACTIVE   |",
                                "| ef7382d2-8009-467e-b354-44b46a384980 | test-2 | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ osism manage compute migrate --yes --target testbed-node-3 testbed-node-4",
                                "\u001b[32m2026-03-21 01:24:16\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-21 01:24:16\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-21 01:24:16\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "\u001b[32m2026-03-21 01:24:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 4740b0e6-2766-4b54-a663-63e0e558e778\u001b[0m",
                                "\u001b[32m2026-03-21 01:24:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:24:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:24:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:24:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:24:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:24:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:24:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:24:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:24:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:24:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:24:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:24:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:25:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-03-21 01:25:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 56e68f34-0e54-4c15-a73d-f764465f4df9\u001b[0m",
                                "\u001b[32m2026-03-21 01:25:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:25:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:25:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:25:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:25:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:25:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:25:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:25:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:25:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:25:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) completed with status ACTIVE\u001b[0m",
                                "+ compute_list",
                                "+ osism manage compute list testbed-node-3",
                                "\u001b[32m2026-03-21 01:25:33\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-21 01:25:33\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-21 01:25:33\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| 4740b0e6-2766-4b54-a663-63e0e558e778 | test   | ACTIVE   |",
                                "| 56e68f34-0e54-4c15-a73d-f764465f4df9 | test-1 | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ osism manage compute list testbed-node-4",
                                "\u001b[32m2026-03-21 01:25:36\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-21 01:25:36\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-21 01:25:36\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ osism manage compute list testbed-node-5",
                                "\u001b[32m2026-03-21 01:25:39\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-21 01:25:39\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-21 01:25:39\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| e341f1ff-bd52-468b-8137-05723da11652 | test-4 | ACTIVE   |",
                                "| a219ecb7-4093-44bb-b939-a453113dfec2 | test-3 | ACTIVE   |",
                                "| ef7382d2-8009-467e-b354-44b46a384980 | test-2 | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ server_ping",
                                "++ tr -d '\\r'",
                                "++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.103",
                                "PING 192.168.112.103 (192.168.112.103) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.103: icmp_seq=1 ttl=63 time=3.75 ms",
                                "64 bytes from 192.168.112.103: icmp_seq=2 ttl=63 time=1.56 ms",
                                "64 bytes from 192.168.112.103: icmp_seq=3 ttl=63 time=1.26 ms",
                                "",
                                "--- 192.168.112.103 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2004ms",
                                "rtt min/avg/max/mdev = 1.260/2.188/3.749/1.110 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.117",
                                "PING 192.168.112.117 (192.168.112.117) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.117: icmp_seq=1 ttl=63 time=5.35 ms",
                                "64 bytes from 192.168.112.117: icmp_seq=2 ttl=63 time=1.28 ms",
                                "64 bytes from 192.168.112.117: icmp_seq=3 ttl=63 time=1.24 ms",
                                "",
                                "--- 192.168.112.117 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.235/2.620/5.348/1.928 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.150",
                                "PING 192.168.112.150 (192.168.112.150) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.150: icmp_seq=1 ttl=63 time=3.55 ms",
                                "64 bytes from 192.168.112.150: icmp_seq=2 ttl=63 time=1.52 ms",
                                "64 bytes from 192.168.112.150: icmp_seq=3 ttl=63 time=1.15 ms",
                                "",
                                "--- 192.168.112.150 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2004ms",
                                "rtt min/avg/max/mdev = 1.152/2.073/3.546/1.052 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.145",
                                "PING 192.168.112.145 (192.168.112.145) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.145: icmp_seq=1 ttl=63 time=3.12 ms",
                                "64 bytes from 192.168.112.145: icmp_seq=2 ttl=63 time=1.37 ms",
                                "64 bytes from 192.168.112.145: icmp_seq=3 ttl=63 time=1.14 ms",
                                "",
                                "--- 192.168.112.145 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.137/1.875/3.121/0.885 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.200",
                                "PING 192.168.112.200 (192.168.112.200) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.200: icmp_seq=1 ttl=63 time=3.30 ms",
                                "64 bytes from 192.168.112.200: icmp_seq=2 ttl=63 time=1.38 ms",
                                "64 bytes from 192.168.112.200: icmp_seq=3 ttl=63 time=1.13 ms",
                                "",
                                "--- 192.168.112.200 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.132/1.934/3.295/0.967 ms",
                                "+ osism manage compute migrate --yes --target testbed-node-3 testbed-node-5",
                                "\u001b[32m2026-03-21 01:25:55\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-21 01:25:55\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-21 01:25:55\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "\u001b[32m2026-03-21 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server e341f1ff-bd52-468b-8137-05723da11652\u001b[0m",
                                "\u001b[32m2026-03-21 01:26:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:26:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:26:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:26:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:26:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:26:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:26:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:26:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:26:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-03-21 01:26:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server a219ecb7-4093-44bb-b939-a453113dfec2\u001b[0m",
                                "\u001b[32m2026-03-21 01:26:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:26:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:26:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:26:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:26:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:26:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:26:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:26:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:26:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-03-21 01:26:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server ef7382d2-8009-467e-b354-44b46a384980\u001b[0m",
                                "\u001b[32m2026-03-21 01:27:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:27:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:27:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:27:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:27:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:27:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:27:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:27:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:27:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) completed with status ACTIVE\u001b[0m",
                                "+ compute_list",
                                "+ osism manage compute list testbed-node-3",
                                "\u001b[32m2026-03-21 01:27:29\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-21 01:27:29\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-21 01:27:29\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| e341f1ff-bd52-468b-8137-05723da11652 | test-4 | ACTIVE   |",
                                "| a219ecb7-4093-44bb-b939-a453113dfec2 | test-3 | ACTIVE   |",
                                "| ef7382d2-8009-467e-b354-44b46a384980 | test-2 | ACTIVE   |",
                                "| 4740b0e6-2766-4b54-a663-63e0e558e778 | test   | ACTIVE   |",
                                "| 56e68f34-0e54-4c15-a73d-f764465f4df9 | test-1 | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ osism manage compute list testbed-node-4",
                                "\u001b[32m2026-03-21 01:27:33\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-21 01:27:33\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-21 01:27:33\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ osism manage compute list testbed-node-5",
                                "\u001b[32m2026-03-21 01:27:35\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-21 01:27:35\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-21 01:27:35\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ server_ping",
                                "++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'",
                                "++ tr -d '\\r'",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.103",
                                "PING 192.168.112.103 (192.168.112.103) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.103: icmp_seq=1 ttl=63 time=3.75 ms",
                                "64 bytes from 192.168.112.103: icmp_seq=2 ttl=63 time=1.62 ms",
                                "64 bytes from 192.168.112.103: icmp_seq=3 ttl=63 time=1.97 ms",
                                "",
                                "--- 192.168.112.103 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.620/2.447/3.754/0.934 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.117",
                                "PING 192.168.112.117 (192.168.112.117) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.117: icmp_seq=1 ttl=63 time=10.7 ms",
                                "64 bytes from 192.168.112.117: icmp_seq=2 ttl=63 time=2.08 ms",
                                "64 bytes from 192.168.112.117: icmp_seq=3 ttl=63 time=1.99 ms",
                                "",
                                "--- 192.168.112.117 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.985/4.911/10.667/4.070 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.150",
                                "PING 192.168.112.150 (192.168.112.150) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.150: icmp_seq=1 ttl=63 time=6.27 ms",
                                "64 bytes from 192.168.112.150: icmp_seq=2 ttl=63 time=2.11 ms",
                                "64 bytes from 192.168.112.150: icmp_seq=3 ttl=63 time=1.55 ms",
                                "",
                                "--- 192.168.112.150 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2002ms",
                                "rtt min/avg/max/mdev = 1.550/3.307/6.267/2.104 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.145",
                                "PING 192.168.112.145 (192.168.112.145) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.145: icmp_seq=1 ttl=63 time=7.33 ms",
                                "64 bytes from 192.168.112.145: icmp_seq=2 ttl=63 time=2.18 ms",
                                "64 bytes from 192.168.112.145: icmp_seq=3 ttl=63 time=1.83 ms",
                                "",
                                "--- 192.168.112.145 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.833/3.779/7.325/2.510 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.200",
                                "PING 192.168.112.200 (192.168.112.200) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.200: icmp_seq=1 ttl=63 time=5.52 ms",
                                "64 bytes from 192.168.112.200: icmp_seq=2 ttl=63 time=1.89 ms",
                                "64 bytes from 192.168.112.200: icmp_seq=3 ttl=63 time=1.42 ms",
                                "",
                                "--- 192.168.112.200 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.419/2.940/5.515/1.830 ms",
                                "+ osism manage compute migrate --yes --target testbed-node-4 testbed-node-3",
                                "\u001b[32m2026-03-21 01:27:51\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-21 01:27:51\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-21 01:27:51\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "\u001b[32m2026-03-21 01:27:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server e341f1ff-bd52-468b-8137-05723da11652\u001b[0m",
                                "\u001b[32m2026-03-21 01:28:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:28:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:28:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:28:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:28:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:28:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:28:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:28:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:28:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-03-21 01:28:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server a219ecb7-4093-44bb-b939-a453113dfec2\u001b[0m",
                                "\u001b[32m2026-03-21 01:28:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:28:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:28:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:28:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:28:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:28:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:28:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:28:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:28:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-03-21 01:28:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server ef7382d2-8009-467e-b354-44b46a384980\u001b[0m",
                                "\u001b[32m2026-03-21 01:29:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:29:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:29:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:29:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:29:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:29:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:29:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:29:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:29:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-03-21 01:29:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 4740b0e6-2766-4b54-a663-63e0e558e778\u001b[0m",
                                "\u001b[32m2026-03-21 01:29:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:29:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:29:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:29:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:29:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:29:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:29:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:29:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:29:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:29:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-03-21 01:29:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 56e68f34-0e54-4c15-a73d-f764465f4df9\u001b[0m",
                                "\u001b[32m2026-03-21 01:30:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:30:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:30:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:30:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:30:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:30:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:30:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:30:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:30:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) completed with status ACTIVE\u001b[0m",
                                "+ compute_list",
                                "+ osism manage compute list testbed-node-3",
                                "\u001b[32m2026-03-21 01:30:25\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-21 01:30:25\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-21 01:30:25\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ osism manage compute list testbed-node-4",
                                "\u001b[32m2026-03-21 01:30:28\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-21 01:30:28\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-21 01:30:28\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| e341f1ff-bd52-468b-8137-05723da11652 | test-4 | ACTIVE   |",
                                "| a219ecb7-4093-44bb-b939-a453113dfec2 | test-3 | ACTIVE   |",
                                "| ef7382d2-8009-467e-b354-44b46a384980 | test-2 | ACTIVE   |",
                                "| 4740b0e6-2766-4b54-a663-63e0e558e778 | test   | ACTIVE   |",
                                "| 56e68f34-0e54-4c15-a73d-f764465f4df9 | test-1 | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ osism manage compute list testbed-node-5",
                                "\u001b[32m2026-03-21 01:30:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-21 01:30:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-21 01:30:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ server_ping",
                                "++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'",
                                "++ tr -d '\\r'",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.103",
                                "PING 192.168.112.103 (192.168.112.103) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.103: icmp_seq=1 ttl=63 time=9.55 ms",
                                "64 bytes from 192.168.112.103: icmp_seq=2 ttl=63 time=2.38 ms",
                                "64 bytes from 192.168.112.103: icmp_seq=3 ttl=63 time=1.70 ms",
                                "",
                                "--- 192.168.112.103 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.701/4.542/9.551/3.552 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.117",
                                "PING 192.168.112.117 (192.168.112.117) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.117: icmp_seq=1 ttl=63 time=7.05 ms",
                                "64 bytes from 192.168.112.117: icmp_seq=2 ttl=63 time=2.38 ms",
                                "64 bytes from 192.168.112.117: icmp_seq=3 ttl=63 time=1.96 ms",
                                "",
                                "--- 192.168.112.117 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.964/3.799/7.054/2.307 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.150",
                                "PING 192.168.112.150 (192.168.112.150) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.150: icmp_seq=1 ttl=63 time=6.13 ms",
                                "64 bytes from 192.168.112.150: icmp_seq=2 ttl=63 time=2.70 ms",
                                "64 bytes from 192.168.112.150: icmp_seq=3 ttl=63 time=1.83 ms",
                                "",
                                "--- 192.168.112.150 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.834/3.554/6.126/1.852 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.145",
                                "PING 192.168.112.145 (192.168.112.145) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.145: icmp_seq=1 ttl=63 time=4.76 ms",
                                "64 bytes from 192.168.112.145: icmp_seq=2 ttl=63 time=2.34 ms",
                                "64 bytes from 192.168.112.145: icmp_seq=3 ttl=63 time=1.90 ms",
                                "",
                                "--- 192.168.112.145 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2004ms",
                                "rtt min/avg/max/mdev = 1.898/2.999/4.761/1.258 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.200",
                                "PING 192.168.112.200 (192.168.112.200) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.200: icmp_seq=1 ttl=63 time=4.50 ms",
                                "64 bytes from 192.168.112.200: icmp_seq=2 ttl=63 time=1.99 ms",
                                "64 bytes from 192.168.112.200: icmp_seq=3 ttl=63 time=1.66 ms",
                                "",
                                "--- 192.168.112.200 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2002ms",
                                "rtt min/avg/max/mdev = 1.655/2.715/4.499/1.268 ms",
                                "+ osism manage compute migrate --yes --target testbed-node-5 testbed-node-4",
                                "\u001b[32m2026-03-21 01:30:48\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-21 01:30:48\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-21 01:30:48\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "\u001b[32m2026-03-21 01:30:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server e341f1ff-bd52-468b-8137-05723da11652\u001b[0m",
                                "\u001b[32m2026-03-21 01:31:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:31:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:31:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:31:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:31:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:31:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:31:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:31:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:31:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e341f1ff-bd52-468b-8137-05723da11652 (test-4) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-03-21 01:31:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server a219ecb7-4093-44bb-b939-a453113dfec2\u001b[0m",
                                "\u001b[32m2026-03-21 01:31:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:31:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:31:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:31:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:31:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:31:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:31:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:31:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:31:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of a219ecb7-4093-44bb-b939-a453113dfec2 (test-3) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-03-21 01:31:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server ef7382d2-8009-467e-b354-44b46a384980\u001b[0m",
                                "\u001b[32m2026-03-21 01:31:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:31:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:32:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:32:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:32:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:32:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:32:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:32:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:32:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of ef7382d2-8009-467e-b354-44b46a384980 (test-2) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-03-21 01:32:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 4740b0e6-2766-4b54-a663-63e0e558e778\u001b[0m",
                                "\u001b[32m2026-03-21 01:32:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:32:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:32:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:32:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:32:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:32:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:32:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:32:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:32:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:32:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:32:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 4740b0e6-2766-4b54-a663-63e0e558e778 (test) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-03-21 01:32:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 56e68f34-0e54-4c15-a73d-f764465f4df9\u001b[0m",
                                "\u001b[32m2026-03-21 01:32:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:33:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:33:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:33:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:33:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:33:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:33:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:33:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-21 01:33:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 56e68f34-0e54-4c15-a73d-f764465f4df9 (test-1) completed with status ACTIVE\u001b[0m",
                                "+ compute_list",
                                "+ osism manage compute list testbed-node-3",
                                "\u001b[32m2026-03-21 01:33:18\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-21 01:33:18\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-21 01:33:18\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ osism manage compute list testbed-node-4",
                                "\u001b[32m2026-03-21 01:33:21\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-21 01:33:21\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-21 01:33:21\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ osism manage compute list testbed-node-5",
                                "\u001b[32m2026-03-21 01:33:23\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-21 01:33:23\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-21 01:33:23\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| e341f1ff-bd52-468b-8137-05723da11652 | test-4 | ACTIVE   |",
                                "| a219ecb7-4093-44bb-b939-a453113dfec2 | test-3 | ACTIVE   |",
                                "| ef7382d2-8009-467e-b354-44b46a384980 | test-2 | ACTIVE   |",
                                "| 4740b0e6-2766-4b54-a663-63e0e558e778 | test   | ACTIVE   |",
                                "| 56e68f34-0e54-4c15-a73d-f764465f4df9 | test-1 | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ server_ping",
                                "++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'",
                                "++ tr -d '\\r'",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.103",
                                "PING 192.168.112.103 (192.168.112.103) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.103: icmp_seq=1 ttl=63 time=7.19 ms",
                                "64 bytes from 192.168.112.103: icmp_seq=2 ttl=63 time=2.12 ms",
                                "64 bytes from 192.168.112.103: icmp_seq=3 ttl=63 time=1.71 ms",
                                "",
                                "--- 192.168.112.103 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.712/3.672/7.189/2.491 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.117",
                                "PING 192.168.112.117 (192.168.112.117) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.117: icmp_seq=1 ttl=63 time=8.33 ms",
                                "64 bytes from 192.168.112.117: icmp_seq=2 ttl=63 time=1.98 ms",
                                "64 bytes from 192.168.112.117: icmp_seq=3 ttl=63 time=1.75 ms",
                                "",
                                "--- 192.168.112.117 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.752/4.022/8.331/3.048 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.150",
                                "PING 192.168.112.150 (192.168.112.150) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.150: icmp_seq=1 ttl=63 time=4.53 ms",
                                "64 bytes from 192.168.112.150: icmp_seq=2 ttl=63 time=1.60 ms",
                                "64 bytes from 192.168.112.150: icmp_seq=3 ttl=63 time=1.76 ms",
                                "",
                                "--- 192.168.112.150 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2004ms",
                                "rtt min/avg/max/mdev = 1.597/2.629/4.532/1.347 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.145",
                                "PING 192.168.112.145 (192.168.112.145) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.145: icmp_seq=1 ttl=63 time=4.75 ms",
                                "64 bytes from 192.168.112.145: icmp_seq=2 ttl=63 time=2.21 ms",
                                "64 bytes from 192.168.112.145: icmp_seq=3 ttl=63 time=1.82 ms",
                                "",
                                "--- 192.168.112.145 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.820/2.928/4.753/1.300 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.200",
                                "PING 192.168.112.200 (192.168.112.200) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.200: icmp_seq=1 ttl=63 time=4.60 ms",
                                "64 bytes from 192.168.112.200: icmp_seq=2 ttl=63 time=2.18 ms",
                                "64 bytes from 192.168.112.200: icmp_seq=3 ttl=63 time=1.79 ms",
                                "",
                                "--- 192.168.112.200 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.786/2.853/4.599/1.244 ms"
                            ],
                            "zuul_log_id": "0242ac1f-6606-9900-9c98-00000000001e-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:33:38.645687Z",
                            "start": "2026-03-21T01:15:16.294352Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-00000000001e",
                        "name": "Run checks"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed04/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.193.136",
                                "/opt/configuration/scripts/check/302-openstack-with-tempest.sh"
                            ],
                            "delta": "0:01:55.076432",
                            "end": "2026-03-21 01:35:34.106338",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed04/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.193.136 /opt/configuration/scripts/check/302-openstack-with-tempest.sh",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-9900-9c98-00000000001f-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-21 01:33:39.029906",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "+ set -e\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=true\n++ TEMPEST=true\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n\n# Tempest\n\n++ CEPH_STACK=ceph-ansible\n+ echo\n+ echo '# Tempest'\n+ echo\n+ [[ ! -e /opt/tempest ]]\n+ osism apply tempest --skip-tags run-tempest\n\u001b[32m2026-03-21 01:33:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of tempest.\u001b[0m\n\u001b[32m2026-03-21 01:33:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ea4b3273-7131-4c96-9be2-eaf43edae2f5 (tempest) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-21 01:33:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task ea4b3273-7131-4c96-9be2-eaf43edae2f5 (tempest) has been started and output is visible here.\u001b[0m\n\nPLAY [Run tempest] *************************************************************\n\nTASK [osism.validations.tempest : Create tempest workdir] **********************\nSaturday 21 March 2026  01:33:54 +0000 (0:00:00.321)       0:00:00.321 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Copy tempest wrapper script] *****************\nSaturday 21 March 2026  01:33:55 +0000 (0:00:01.060)       0:00:01.381 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Check for existing tempest initialisation] ***\nSaturday 21 March 2026  01:33:56 +0000 (0:00:01.206)       0:00:02.588 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Init tempest] ********************************\nSaturday 21 March 2026  01:33:57 +0000 (0:00:00.421)       0:00:03.009 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Resolve image IDs] ***************************\nSaturday 21 March 2026  01:34:17 +0000 (0:00:20.421)       0:00:23.431 ******** \n\u001b[0;32mok: [testbed-manager -> localhost] => (item=Cirros 0.6.3)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost] => (item=Cirros 0.6.2)\u001b[0m\n\nTASK [osism.validations.tempest : Assert images have been resolved] ************\nSaturday 21 March 2026  01:34:25 +0000 (0:00:08.020)       0:00:31.452 ******** \n\u001b[0;32mok: [testbed-manager] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [osism.validations.tempest : Get auth token] ******************************\nSaturday 21 March 2026  01:34:25 +0000 (0:00:00.156)       0:00:31.608 ******** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Get endpoint catalog] ************************\nSaturday 21 March 2026  01:34:29 +0000 (0:00:03.617)       0:00:35.225 ******** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Get service catalog] *************************\nSaturday 21 March 2026  01:34:31 +0000 (0:00:01.881)       0:00:37.106 ******** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Register img_file name] **********************\nSaturday 21 March 2026  01:34:34 +0000 (0:00:03.613)       0:00:40.719 ******** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Download img_file from image_ref] ************\nSaturday 21 March 2026  01:34:35 +0000 (0:00:00.188)       0:00:40.908 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Install qemu-utils package] ******************\nSaturday 21 March 2026  01:34:37 +0000 (0:00:02.864)       0:00:43.773 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Convert img_file to qcow2 format] ************\nSaturday 21 March 2026  01:34:46 +0000 (0:00:08.837)       0:00:52.610 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Get network API extensions] ******************\nSaturday 21 March 2026  01:34:47 +0000 (0:00:00.669)       0:00:53.280 ******** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Revoke token] ********************************\nSaturday 21 March 2026  01:34:48 +0000 (0:00:01.510)       0:00:54.790 ******** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Set fact for config option api_extensions] ***\nSaturday 21 March 2026  01:34:50 +0000 (0:00:01.518)       0:00:56.308 ******** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Set fact for config option img_file] *********\nSaturday 21 March 2026  01:34:50 +0000 (0:00:00.187)       0:00:56.496 ******** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Resolve floating network ID] *****************\nSaturday 21 March 2026  01:34:50 +0000 (0:00:00.355)       0:00:56.851 ******** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Assert floating network id has been resolved] ***\nSaturday 21 March 2026  01:34:54 +0000 (0:00:03.903)       0:01:00.754 ******** \n\u001b[0;32mok: [testbed-manager -> localhost] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [osism.validations.tempest : Resolve flavor IDs] **************************\nSaturday 21 March 2026  01:34:55 +0000 (0:00:00.177)       0:01:00.932 ******** \n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'tempest-1', 'vcpus': 1, 'ram': 1024, 'disk': 1}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'tempest-2', 'vcpus': 2, 'ram': 2048, 'disk': 2}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Assert flavors have been resolved] ***********\nSaturday 21 March 2026  01:34:55 +0000 (0:00:00.204)       0:01:01.136 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Get stats of exclude list] *******************\nSaturday 21 March 2026  01:34:55 +0000 (0:00:00.164)       0:01:01.301 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Copy exclude list] ***************************\nSaturday 21 March 2026  01:34:55 +0000 (0:00:00.452)       0:01:01.754 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Get stats of include list] *******************\nSaturday 21 March 2026  01:34:56 +0000 (0:00:00.889)       0:01:02.644 ******** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Copy include list] ***************************\nSaturday 21 March 2026  01:34:57 +0000 (0:00:00.418)       0:01:03.062 ******** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Create tempest flavors] **********************\nSaturday 21 March 2026  01:34:57 +0000 (0:00:00.308)       0:01:03.371 ******** \n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item={'name': 'tempest-1', 'vcpus': 1, 'ram': 1024, 'disk': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item={'name': 'tempest-2', 'vcpus': 2, 'ram': 2048, 'disk': 2})\u001b[0m\n\nTASK [osism.validations.tempest : Copy tempest.conf file] **********************\nSaturday 21 March 2026  01:35:05 +0000 (0:00:07.748)       0:01:11.120 ******** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nSaturday 21 March 2026  01:35:06 +0000 (0:00:01.021)       0:01:12.141 ******** \n=============================================================================== \nosism.validations.tempest : Init tempest ------------------------------- 20.42s\nosism.validations.tempest : Install qemu-utils package ------------------ 8.84s\nosism.validations.tempest : Resolve image IDs --------------------------- 8.02s\nosism.validations.tempest : Create tempest flavors ---------------------- 7.75s\nosism.validations.tempest : Resolve floating network ID ----------------- 3.90s\nosism.validations.tempest : Get auth token ------------------------------ 3.62s\nosism.validations.tempest : Get service catalog ------------------------- 3.61s\nosism.validations.tempest : Download img_file from image_ref ------------ 2.86s\nosism.validations.tempest : Get endpoint catalog ------------------------ 1.88s\nosism.validations.tempest : Revoke token -------------------------------- 1.52s\nosism.validations.tempest : Get network API extensions ------------------ 1.51s\nosism.validations.tempest : Copy tempest wrapper script ----------------- 1.21s\nosism.validations.tempest : Create tempest workdir ---------------------- 1.06s\nosism.validations.tempest : Copy tempest.conf file ---------------------- 1.02s\nosism.validations.tempest : Copy exclude list --------------------------- 0.89s\nosism.validations.tempest : Convert img_file to qcow2 format ------------ 0.67s\nosism.validations.tempest : Get stats of exclude list ------------------- 0.45s\nosism.validations.tempest : Check for existing tempest initialisation --- 0.42s\nosism.validations.tempest : Get stats of include list ------------------- 0.42s\nosism.validations.tempest : Set fact for config option img_file --------- 0.36s\n+ sed -i '/log_dir =/d' /opt/tempest/etc/tempest.conf\n+ sed -i '/log_file =/d' /opt/tempest/etc/tempest.conf\n\n## IDENTITY (API)\n\n+ [[ false == \\t\\r\\u\\e ]]\n+ echo\n+ echo '## IDENTITY (API)'\n+ echo\n+ _tempest tempest.api.identity.v3\n+ local regex=tempest.api.identity.v3\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.identity.v3 --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260321-0135.log\ntempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.identity.v3 --concurrency 16' is not a tempest command. See 'tempest --help'.\nDid you mean one of these?\n  help\n  init\n\n## IMAGE (API)\n\n+ echo\n+ echo '## IMAGE (API)'\n+ echo\n+ _tempest tempest.api.image.v2\n+ local regex=tempest.api.image.v2\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.image.v2 --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260321-0135.log\ntempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.image.v2 --concurrency 16' is not a tempest command. See 'tempest --help'.\nDid you mean one of these?\n  help\n  init\n\n## NETWORK (API)\n\n+ echo\n+ echo '## NETWORK (API)'\n+ echo\n+ _tempest tempest.api.network\n+ local regex=tempest.api.network\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.network --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260321-0135.log\ntempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.network --concurrency 16' is not a tempest command. See 'tempest --help'.\nDid you mean one of these?\n  help\n  init\n\n## VOLUME (API)\n\n+ echo\n+ echo '## VOLUME (API)'\n+ echo\n+ _tempest tempest.api.volume\n+ local regex=tempest.api.volume\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.volume --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260321-0135.log\ntempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.volume --concurrency 16' is not a tempest command. See 'tempest --help'.\nDid you mean one of these?\n  help\n  init\n\n## COMPUTE (API)\n\n+ echo\n+ echo '## COMPUTE (API)'\n+ echo\n+ _tempest tempest.api.compute\n+ local regex=tempest.api.compute\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.compute --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260321-0135.log\ntempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.compute --concurrency 16' is not a tempest command. See 'tempest --help'.\nDid you mean one of these?\n  help\n  init\n\n## DNS (API)\n\n+ echo\n+ echo '## DNS (API)'\n+ echo\n+ _tempest designate_tempest_plugin.tests.api.v2\n+ local regex=designate_tempest_plugin.tests.api.v2\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex designate_tempest_plugin.tests.api.v2 --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260321-0135.log\ntempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex designate_tempest_plugin.tests.api.v2 --concurrency 16' is not a tempest command. See 'tempest --help'.\nDid you mean one of these?\n  help\n  init\n\n## OBJECT-STORE (API)\n\n+ echo\n+ echo '## OBJECT-STORE (API)'\n+ echo\n+ _tempest tempest.api.object_storage\n+ local regex=tempest.api.object_storage\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.object_storage --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260321-0135.log\ntempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.object_storage --concurrency 16' is not a tempest command. See 'tempest --help'.\nDid you mean one of these?\n  help\n  init",
                            "stdout_lines": [
                                "+ set -e",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=true",
                                "++ TEMPEST=true",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.193.136",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "",
                                "# Tempest",
                                "",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ echo",
                                "+ echo '# Tempest'",
                                "+ echo",
                                "+ [[ ! -e /opt/tempest ]]",
                                "+ osism apply tempest --skip-tags run-tempest",
                                "\u001b[32m2026-03-21 01:33:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of tempest.\u001b[0m",
                                "\u001b[32m2026-03-21 01:33:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ea4b3273-7131-4c96-9be2-eaf43edae2f5 (tempest) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-21 01:33:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task ea4b3273-7131-4c96-9be2-eaf43edae2f5 (tempest) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Run tempest] *************************************************************",
                                "",
                                "TASK [osism.validations.tempest : Create tempest workdir] **********************",
                                "Saturday 21 March 2026  01:33:54 +0000 (0:00:00.321)       0:00:00.321 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Copy tempest wrapper script] *****************",
                                "Saturday 21 March 2026  01:33:55 +0000 (0:00:01.060)       0:00:01.381 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Check for existing tempest initialisation] ***",
                                "Saturday 21 March 2026  01:33:56 +0000 (0:00:01.206)       0:00:02.588 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Init tempest] ********************************",
                                "Saturday 21 March 2026  01:33:57 +0000 (0:00:00.421)       0:00:03.009 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Resolve image IDs] ***************************",
                                "Saturday 21 March 2026  01:34:17 +0000 (0:00:20.421)       0:00:23.431 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost] => (item=Cirros 0.6.3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost] => (item=Cirros 0.6.2)\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Assert images have been resolved] ************",
                                "Saturday 21 March 2026  01:34:25 +0000 (0:00:08.020)       0:00:31.452 ******** ",
                                "\u001b[0;32mok: [testbed-manager] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Get auth token] ******************************",
                                "Saturday 21 March 2026  01:34:25 +0000 (0:00:00.156)       0:00:31.608 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Get endpoint catalog] ************************",
                                "Saturday 21 March 2026  01:34:29 +0000 (0:00:03.617)       0:00:35.225 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Get service catalog] *************************",
                                "Saturday 21 March 2026  01:34:31 +0000 (0:00:01.881)       0:00:37.106 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Register img_file name] **********************",
                                "Saturday 21 March 2026  01:34:34 +0000 (0:00:03.613)       0:00:40.719 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Download img_file from image_ref] ************",
                                "Saturday 21 March 2026  01:34:35 +0000 (0:00:00.188)       0:00:40.908 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Install qemu-utils package] ******************",
                                "Saturday 21 March 2026  01:34:37 +0000 (0:00:02.864)       0:00:43.773 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Convert img_file to qcow2 format] ************",
                                "Saturday 21 March 2026  01:34:46 +0000 (0:00:08.837)       0:00:52.610 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Get network API extensions] ******************",
                                "Saturday 21 March 2026  01:34:47 +0000 (0:00:00.669)       0:00:53.280 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Revoke token] ********************************",
                                "Saturday 21 March 2026  01:34:48 +0000 (0:00:01.510)       0:00:54.790 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Set fact for config option api_extensions] ***",
                                "Saturday 21 March 2026  01:34:50 +0000 (0:00:01.518)       0:00:56.308 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Set fact for config option img_file] *********",
                                "Saturday 21 March 2026  01:34:50 +0000 (0:00:00.187)       0:00:56.496 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Resolve floating network ID] *****************",
                                "Saturday 21 March 2026  01:34:50 +0000 (0:00:00.355)       0:00:56.851 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Assert floating network id has been resolved] ***",
                                "Saturday 21 March 2026  01:34:54 +0000 (0:00:03.903)       0:01:00.754 ******** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Resolve flavor IDs] **************************",
                                "Saturday 21 March 2026  01:34:55 +0000 (0:00:00.177)       0:01:00.932 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'tempest-1', 'vcpus': 1, 'ram': 1024, 'disk': 1}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'tempest-2', 'vcpus': 2, 'ram': 2048, 'disk': 2}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Assert flavors have been resolved] ***********",
                                "Saturday 21 March 2026  01:34:55 +0000 (0:00:00.204)       0:01:01.136 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Get stats of exclude list] *******************",
                                "Saturday 21 March 2026  01:34:55 +0000 (0:00:00.164)       0:01:01.301 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Copy exclude list] ***************************",
                                "Saturday 21 March 2026  01:34:55 +0000 (0:00:00.452)       0:01:01.754 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Get stats of include list] *******************",
                                "Saturday 21 March 2026  01:34:56 +0000 (0:00:00.889)       0:01:02.644 ******** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Copy include list] ***************************",
                                "Saturday 21 March 2026  01:34:57 +0000 (0:00:00.418)       0:01:03.062 ******** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Create tempest flavors] **********************",
                                "Saturday 21 March 2026  01:34:57 +0000 (0:00:00.308)       0:01:03.371 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item={'name': 'tempest-1', 'vcpus': 1, 'ram': 1024, 'disk': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item={'name': 'tempest-2', 'vcpus': 2, 'ram': 2048, 'disk': 2})\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Copy tempest.conf file] **********************",
                                "Saturday 21 March 2026  01:35:05 +0000 (0:00:07.748)       0:01:11.120 ******** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Saturday 21 March 2026  01:35:06 +0000 (0:00:01.021)       0:01:12.141 ******** ",
                                "=============================================================================== ",
                                "osism.validations.tempest : Init tempest ------------------------------- 20.42s",
                                "osism.validations.tempest : Install qemu-utils package ------------------ 8.84s",
                                "osism.validations.tempest : Resolve image IDs --------------------------- 8.02s",
                                "osism.validations.tempest : Create tempest flavors ---------------------- 7.75s",
                                "osism.validations.tempest : Resolve floating network ID ----------------- 3.90s",
                                "osism.validations.tempest : Get auth token ------------------------------ 3.62s",
                                "osism.validations.tempest : Get service catalog ------------------------- 3.61s",
                                "osism.validations.tempest : Download img_file from image_ref ------------ 2.86s",
                                "osism.validations.tempest : Get endpoint catalog ------------------------ 1.88s",
                                "osism.validations.tempest : Revoke token -------------------------------- 1.52s",
                                "osism.validations.tempest : Get network API extensions ------------------ 1.51s",
                                "osism.validations.tempest : Copy tempest wrapper script ----------------- 1.21s",
                                "osism.validations.tempest : Create tempest workdir ---------------------- 1.06s",
                                "osism.validations.tempest : Copy tempest.conf file ---------------------- 1.02s",
                                "osism.validations.tempest : Copy exclude list --------------------------- 0.89s",
                                "osism.validations.tempest : Convert img_file to qcow2 format ------------ 0.67s",
                                "osism.validations.tempest : Get stats of exclude list ------------------- 0.45s",
                                "osism.validations.tempest : Check for existing tempest initialisation --- 0.42s",
                                "osism.validations.tempest : Get stats of include list ------------------- 0.42s",
                                "osism.validations.tempest : Set fact for config option img_file --------- 0.36s",
                                "+ sed -i '/log_dir =/d' /opt/tempest/etc/tempest.conf",
                                "+ sed -i '/log_file =/d' /opt/tempest/etc/tempest.conf",
                                "",
                                "## IDENTITY (API)",
                                "",
                                "+ [[ false == \\t\\r\\u\\e ]]",
                                "+ echo",
                                "+ echo '## IDENTITY (API)'",
                                "+ echo",
                                "+ _tempest tempest.api.identity.v3",
                                "+ local regex=tempest.api.identity.v3",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.identity.v3 --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260321-0135.log",
                                "tempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.identity.v3 --concurrency 16' is not a tempest command. See 'tempest --help'.",
                                "Did you mean one of these?",
                                "  help",
                                "  init",
                                "",
                                "## IMAGE (API)",
                                "",
                                "+ echo",
                                "+ echo '## IMAGE (API)'",
                                "+ echo",
                                "+ _tempest tempest.api.image.v2",
                                "+ local regex=tempest.api.image.v2",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.image.v2 --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260321-0135.log",
                                "tempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.image.v2 --concurrency 16' is not a tempest command. See 'tempest --help'.",
                                "Did you mean one of these?",
                                "  help",
                                "  init",
                                "",
                                "## NETWORK (API)",
                                "",
                                "+ echo",
                                "+ echo '## NETWORK (API)'",
                                "+ echo",
                                "+ _tempest tempest.api.network",
                                "+ local regex=tempest.api.network",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.network --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260321-0135.log",
                                "tempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.network --concurrency 16' is not a tempest command. See 'tempest --help'.",
                                "Did you mean one of these?",
                                "  help",
                                "  init",
                                "",
                                "## VOLUME (API)",
                                "",
                                "+ echo",
                                "+ echo '## VOLUME (API)'",
                                "+ echo",
                                "+ _tempest tempest.api.volume",
                                "+ local regex=tempest.api.volume",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.volume --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260321-0135.log",
                                "tempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.volume --concurrency 16' is not a tempest command. See 'tempest --help'.",
                                "Did you mean one of these?",
                                "  help",
                                "  init",
                                "",
                                "## COMPUTE (API)",
                                "",
                                "+ echo",
                                "+ echo '## COMPUTE (API)'",
                                "+ echo",
                                "+ _tempest tempest.api.compute",
                                "+ local regex=tempest.api.compute",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.compute --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260321-0135.log",
                                "tempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.compute --concurrency 16' is not a tempest command. See 'tempest --help'.",
                                "Did you mean one of these?",
                                "  help",
                                "  init",
                                "",
                                "## DNS (API)",
                                "",
                                "+ echo",
                                "+ echo '## DNS (API)'",
                                "+ echo",
                                "+ _tempest designate_tempest_plugin.tests.api.v2",
                                "+ local regex=designate_tempest_plugin.tests.api.v2",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex designate_tempest_plugin.tests.api.v2 --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260321-0135.log",
                                "tempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex designate_tempest_plugin.tests.api.v2 --concurrency 16' is not a tempest command. See 'tempest --help'.",
                                "Did you mean one of these?",
                                "  help",
                                "  init",
                                "",
                                "## OBJECT-STORE (API)",
                                "",
                                "+ echo",
                                "+ echo '## OBJECT-STORE (API)'",
                                "+ echo",
                                "+ _tempest tempest.api.object_storage",
                                "+ local regex=tempest.api.object_storage",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.object_storage --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260321-0135.log",
                                "tempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.object_storage --concurrency 16' is not a tempest command. See 'tempest --help'.",
                                "Did you mean one of these?",
                                "  help",
                                "  init"
                            ],
                            "zuul_log_id": "0242ac1f-6606-9900-9c98-00000000001f-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:35:34.275076Z",
                            "start": "2026-03-21T01:33:38.694785Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-00000000001f",
                        "name": "Run tempest"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": false,
                            "false_condition": "_prometheus_alert_status | bool",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:35:34.829570Z",
                            "start": "2026-03-21T01:35:34.319672Z"
                        },
                        "id": "0242ac1f-6606-9900-9c98-000000000020",
                        "name": "Check prometheus alert status"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 12,
            "failures": 0,
            "ignored": 0,
            "ok": 25,
            "rescued": 0,
            "skipped": 4,
            "unreachable": 0
        }
    },
    "trusted": false
},
{
    "branch": "main",
    "index": "0",
    "phase": "post",
    "playbook": "github.com/osism/testbed/playbooks/post.yml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-03-21T01:35:38.714407Z",
                    "start": "2026-03-21T01:35:35.867838Z"
                },
                "id": "0242ac1f-6606-7058-a22c-000000000002",
                "name": "Post output play"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "stat",
                            "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
                            "changed": false,
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-7058-a22c-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:35:35.929476Z",
                            "start": "2026-03-21T01:35:35.884332Z"
                        },
                        "id": "0242ac1f-6606-7058-a22c-000000000007",
                        "name": "Register sources"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": true,
                            "cmd": [
                                "sudo",
                                "-n",
                                "true"
                            ],
                            "delta": "0:00:00.011324",
                            "end": "2026-03-21 01:35:36.782690",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "sudo -n true",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-7058-a22c-000000000009-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "non-zero return code",
                            "rc": 1,
                            "start": "2026-03-21 01:35:36.771366",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "sudo: a password is required",
                            "stdout_lines": [
                                "sudo: a password is required"
                            ],
                            "zuul_log_id": "0242ac1f-6606-7058-a22c-000000000009-1-orchestrator"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-7058-a22c-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:35:36.973996Z",
                            "start": "2026-03-21T01:35:35.962151Z"
                        },
                        "id": "0242ac1f-6606-7058-a22c-000000000009",
                        "name": "Check sudo"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
                            "changed": false,
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-7058-a22c-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:35:37.015774Z",
                            "start": "2026-03-21T01:35:36.989264Z"
                        },
                        "id": "0242ac1f-6606-7058-a22c-00000000000a",
                        "name": "Set source and destination for files and folders"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "all_sources": []
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-7058-a22c-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:35:37.100437Z",
                            "start": "2026-03-21T01:35:37.031648Z"
                        },
                        "id": "0242ac1f-6606-7058-a22c-00000000000c",
                        "name": "Build a list of source, dest dictionaries"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": false,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_output_dirs",
                                    "changed": false,
                                    "diff": {
                                        "after": {
                                            "path": "/home/zuul-testbed04/zuul-output/docs"
                                        },
                                        "before": {
                                            "path": "/home/zuul-testbed04/zuul-output/docs"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 1005,
                                    "group": "zuul-testbed04",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": "zuul-testbed04",
                                            "path": "/home/zuul-testbed04/zuul-output/docs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "zuul-testbed04",
                                    "path": "/home/zuul-testbed04/zuul-output/docs",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 1004,
                                    "zj_output_dirs": "docs"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dirs",
                                    "changed": false,
                                    "diff": {
                                        "after": {
                                            "path": "/home/zuul-testbed04/zuul-output/artifacts"
                                        },
                                        "before": {
                                            "path": "/home/zuul-testbed04/zuul-output/artifacts"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 1005,
                                    "group": "zuul-testbed04",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": "zuul-testbed04",
                                            "path": "/home/zuul-testbed04/zuul-output/artifacts",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "zuul-testbed04",
                                    "path": "/home/zuul-testbed04/zuul-output/artifacts",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 1004,
                                    "zj_output_dirs": "artifacts"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dirs",
                                    "changed": false,
                                    "diff": {
                                        "after": {
                                            "path": "/home/zuul-testbed04/zuul-output/logs"
                                        },
                                        "before": {
                                            "path": "/home/zuul-testbed04/zuul-output/logs"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 1005,
                                    "group": "zuul-testbed04",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": "zuul-testbed04",
                                            "path": "/home/zuul-testbed04/zuul-output/logs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "zuul-testbed04",
                                    "path": "/home/zuul-testbed04/zuul-output/logs",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 1004,
                                    "zj_output_dirs": "logs"
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-7058-a22c-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:35:38.223355Z",
                            "start": "2026-03-21T01:35:37.109282Z"
                        },
                        "id": "0242ac1f-6606-7058-a22c-00000000000d",
                        "name": "Ensure target folders exist"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": false,
                            "results": [],
                            "skipped": true,
                            "skipped_reason": "No items in the list"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-7058-a22c-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:35:38.255417Z",
                            "start": "2026-03-21T01:35:38.231049Z"
                        },
                        "id": "0242ac1f-6606-7058-a22c-00000000000f",
                        "name": "Copy files and folders to staging folder"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": false,
                            "diff": {
                                "after": {
                                    "path": "/home/zuul-testbed04/zuul-output/logs"
                                },
                                "before": {
                                    "path": "/home/zuul-testbed04/zuul-output/logs"
                                }
                            },
                            "gid": 1005,
                            "group": "zuul-testbed04",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "dest": "/home/zuul-testbed04/zuul-output/logs",
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": "u=rwX,g=rX,o=rX",
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/home/zuul-testbed04/zuul-output/logs",
                                    "recurse": true,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0755",
                            "owner": "zuul-testbed04",
                            "path": "/home/zuul-testbed04/zuul-output/logs",
                            "size": 4096,
                            "state": "directory",
                            "uid": 1004
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-7058-a22c-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:35:38.591471Z",
                            "start": "2026-03-21T01:35:38.267490Z"
                        },
                        "id": "0242ac1f-6606-7058-a22c-000000000011",
                        "name": "Make all log files readable"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "stage_output_renames",
                            "changed": false,
                            "false_condition": "extensions_to_txt is defined and extensions_to_txt is not none",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-7058-a22c-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:35:38.634646Z",
                            "start": "2026-03-21T01:35:38.599334Z"
                        },
                        "id": "0242ac1f-6606-7058-a22c-000000000012",
                        "name": "Rename log files that match extensions_to_txt"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "find",
                            "changed": false,
                            "false_condition": "stage_compress_logs",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-7058-a22c-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:35:38.674606Z",
                            "start": "2026-03-21T01:35:38.649451Z"
                        },
                        "id": "0242ac1f-6606-7058-a22c-000000000014",
                        "name": "Discover log files for compression"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "shell",
                            "changed": false,
                            "false_condition": "stage_compress_logs",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-7058-a22c-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:35:38.714407Z",
                            "start": "2026-03-21T01:35:38.689775Z"
                        },
                        "id": "0242ac1f-6606-7058-a22c-000000000015",
                        "name": "Archive everything from logs"
                    }
                }
            ]
        },
        {
            "play": {
                "duration": {
                    "end": "2026-03-21T01:36:01.424820Z",
                    "start": "2026-03-21T01:35:38.736379Z"
                },
                "id": "0242ac1f-6606-7058-a22c-000000000017",
                "name": "Post cleanup play"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "cloud": "ci"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:35:38.803592Z",
                            "start": "2026-03-21T01:35:38.745635Z"
                        },
                        "id": "0242ac1f-6606-7058-a22c-000000000019",
                        "name": "Set cloud fact (Zuul deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "changed": false,
                            "false_condition": "'nodepool' not in hostvars[groups['all'][0]]",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:35:38.850205Z",
                            "start": "2026-03-21T01:35:38.815176Z"
                        },
                        "id": "0242ac1f-6606-7058-a22c-00000000001a",
                        "name": "Set cloud fact (local deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                            "delta": "0:00:22.122173",
                            "end": "2026-03-21 01:36:01.306738",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed04/src/github.com/osism/testbed/terraform/scripts",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-7058-a22c-00000000001b-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-21 01:35:39.184565",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "2026-03-21 01:35:40 - clean up servers\n2026-03-21 01:35:41 - testbed-manager\n2026-03-21 01:35:41 - testbed-node-3\n2026-03-21 01:35:41 - testbed-node-1\n2026-03-21 01:35:41 - testbed-node-2\n2026-03-21 01:35:41 - testbed-node-0\n2026-03-21 01:35:42 - testbed-node-5\n2026-03-21 01:35:42 - testbed-node-4\n2026-03-21 01:35:42 - clean up keypairs\n2026-03-21 01:35:42 - testbed\n2026-03-21 01:35:42 - wait for servers to be gone\n2026-03-21 01:35:53 - clean up ports\n2026-03-21 01:35:53 - 3f0d3a54-e327-4a03-b440-d8c037e98238\n2026-03-21 01:35:54 - 526a6611-a846-4e7b-a6a1-0efffdc856e4\n2026-03-21 01:35:54 - 6d2237e8-6b42-425e-9866-5df0c0fde3a4\n2026-03-21 01:35:54 - 721c56ff-7560-4029-9a8f-7884ed9b5d65\n2026-03-21 01:35:54 - 761114b3-7304-40c6-96a1-45962f98d109\n2026-03-21 01:35:55 - e3bbeb11-4568-4312-9667-3bd35e5b9212\n2026-03-21 01:35:55 - ec98f5b1-4c59-433a-a2fe-b28505829035\n2026-03-21 01:35:55 - clean up volumes\n2026-03-21 01:35:55 - testbed-volume-4-node-base\n2026-03-21 01:35:55 - testbed-volume-2-node-base\n2026-03-21 01:35:55 - testbed-volume-3-node-base\n2026-03-21 01:35:55 - testbed-volume-5-node-base\n2026-03-21 01:35:56 - testbed-volume-0-node-base\n2026-03-21 01:35:56 - testbed-volume-1-node-base\n2026-03-21 01:35:56 - testbed-volume-manager-base\n2026-03-21 01:35:56 - testbed-volume-8-node-5\n2026-03-21 01:35:56 - testbed-volume-3-node-3\n2026-03-21 01:35:56 - testbed-volume-0-node-3\n2026-03-21 01:35:56 - testbed-volume-5-node-5\n2026-03-21 01:35:56 - testbed-volume-2-node-5\n2026-03-21 01:35:56 - testbed-volume-6-node-3\n2026-03-21 01:35:56 - testbed-volume-1-node-4\n2026-03-21 01:35:56 - testbed-volume-4-node-4\n2026-03-21 01:35:56 - testbed-volume-7-node-4\n2026-03-21 01:35:56 - disconnect routers\n2026-03-21 01:35:56 - testbed\n2026-03-21 01:35:57 - clean up subnets\n2026-03-21 01:35:58 - subnet-testbed-management\n2026-03-21 01:35:58 - clean up networks\n2026-03-21 01:35:58 - net-testbed-management\n2026-03-21 01:35:59 - clean up security groups\n2026-03-21 01:35:59 - testbed-management\n2026-03-21 01:35:59 - testbed-node\n2026-03-21 01:35:59 - clean up floating ips\n2026-03-21 01:35:59 - 81.163.193.136\n2026-03-21 01:35:59 - clean up routers\n2026-03-21 01:35:59 - testbed",
                            "stdout_lines": [
                                "2026-03-21 01:35:40 - clean up servers",
                                "2026-03-21 01:35:41 - testbed-manager",
                                "2026-03-21 01:35:41 - testbed-node-3",
                                "2026-03-21 01:35:41 - testbed-node-1",
                                "2026-03-21 01:35:41 - testbed-node-2",
                                "2026-03-21 01:35:41 - testbed-node-0",
                                "2026-03-21 01:35:42 - testbed-node-5",
                                "2026-03-21 01:35:42 - testbed-node-4",
                                "2026-03-21 01:35:42 - clean up keypairs",
                                "2026-03-21 01:35:42 - testbed",
                                "2026-03-21 01:35:42 - wait for servers to be gone",
                                "2026-03-21 01:35:53 - clean up ports",
                                "2026-03-21 01:35:53 - 3f0d3a54-e327-4a03-b440-d8c037e98238",
                                "2026-03-21 01:35:54 - 526a6611-a846-4e7b-a6a1-0efffdc856e4",
                                "2026-03-21 01:35:54 - 6d2237e8-6b42-425e-9866-5df0c0fde3a4",
                                "2026-03-21 01:35:54 - 721c56ff-7560-4029-9a8f-7884ed9b5d65",
                                "2026-03-21 01:35:54 - 761114b3-7304-40c6-96a1-45962f98d109",
                                "2026-03-21 01:35:55 - e3bbeb11-4568-4312-9667-3bd35e5b9212",
                                "2026-03-21 01:35:55 - ec98f5b1-4c59-433a-a2fe-b28505829035",
                                "2026-03-21 01:35:55 - clean up volumes",
                                "2026-03-21 01:35:55 - testbed-volume-4-node-base",
                                "2026-03-21 01:35:55 - testbed-volume-2-node-base",
                                "2026-03-21 01:35:55 - testbed-volume-3-node-base",
                                "2026-03-21 01:35:55 - testbed-volume-5-node-base",
                                "2026-03-21 01:35:56 - testbed-volume-0-node-base",
                                "2026-03-21 01:35:56 - testbed-volume-1-node-base",
                                "2026-03-21 01:35:56 - testbed-volume-manager-base",
                                "2026-03-21 01:35:56 - testbed-volume-8-node-5",
                                "2026-03-21 01:35:56 - testbed-volume-3-node-3",
                                "2026-03-21 01:35:56 - testbed-volume-0-node-3",
                                "2026-03-21 01:35:56 - testbed-volume-5-node-5",
                                "2026-03-21 01:35:56 - testbed-volume-2-node-5",
                                "2026-03-21 01:35:56 - testbed-volume-6-node-3",
                                "2026-03-21 01:35:56 - testbed-volume-1-node-4",
                                "2026-03-21 01:35:56 - testbed-volume-4-node-4",
                                "2026-03-21 01:35:56 - testbed-volume-7-node-4",
                                "2026-03-21 01:35:56 - disconnect routers",
                                "2026-03-21 01:35:56 - testbed",
                                "2026-03-21 01:35:57 - clean up subnets",
                                "2026-03-21 01:35:58 - subnet-testbed-management",
                                "2026-03-21 01:35:58 - clean up networks",
                                "2026-03-21 01:35:58 - net-testbed-management",
                                "2026-03-21 01:35:59 - clean up security groups",
                                "2026-03-21 01:35:59 - testbed-management",
                                "2026-03-21 01:35:59 - testbed-node",
                                "2026-03-21 01:35:59 - clean up floating ips",
                                "2026-03-21 01:35:59 - 81.163.193.136",
                                "2026-03-21 01:35:59 - clean up routers",
                                "2026-03-21 01:35:59 - testbed"
                            ],
                            "zuul_log_id": "0242ac1f-6606-7058-a22c-00000000001b-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:36:01.424820Z",
                            "start": "2026-03-21T01:35:38.892615Z"
                        },
                        "id": "0242ac1f-6606-7058-a22c-00000000001b",
                        "name": "Clean the cloud environment"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 2,
            "failures": 0,
            "ignored": 0,
            "ok": 6,
            "rescued": 0,
            "skipped": 7,
            "unreachable": 0
        }
    },
    "trusted": false
},
{
    "branch": "main",
    "index": "1",
    "phase": "post",
    "playbook": "github.com/osism/testbed/playbooks/cleanup.yml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-03-21T01:36:05.001563Z",
                    "start": "2026-03-21T01:36:02.300868Z"
                },
                "id": "0242ac1f-6606-c9d8-0c48-000000000002",
                "name": "Cleanup play"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "cloud": "ci"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:36:02.387127Z",
                            "start": "2026-03-21T01:36:02.317190Z"
                        },
                        "id": "0242ac1f-6606-c9d8-0c48-000000000004",
                        "name": "Set cloud fact (Zuul deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "changed": false,
                            "false_condition": "'nodepool' not in hostvars[groups['all'][0]]",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:36:02.443274Z",
                            "start": "2026-03-21T01:36:02.396661Z"
                        },
                        "id": "0242ac1f-6606-c9d8-0c48-000000000005",
                        "name": "Set cloud fact (local deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                            "delta": "0:00:01.413036",
                            "end": "2026-03-21 01:36:04.710834",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed04/src/github.com/osism/testbed/terraform/scripts",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-c9d8-0c48-000000000006-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-21 01:36:03.297798",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "2026-03-21 01:36:03 - clean up servers\n2026-03-21 01:36:04 - clean up keypairs\n2026-03-21 01:36:04 - wait for servers to be gone\n2026-03-21 01:36:04 - clean up ports\n2026-03-21 01:36:04 - clean up volumes\n2026-03-21 01:36:04 - disconnect routers\n2026-03-21 01:36:04 - clean up subnets\n2026-03-21 01:36:04 - clean up networks\n2026-03-21 01:36:04 - clean up security groups\n2026-03-21 01:36:04 - clean up floating ips\n2026-03-21 01:36:04 - clean up routers",
                            "stdout_lines": [
                                "2026-03-21 01:36:03 - clean up servers",
                                "2026-03-21 01:36:04 - clean up keypairs",
                                "2026-03-21 01:36:04 - wait for servers to be gone",
                                "2026-03-21 01:36:04 - clean up ports",
                                "2026-03-21 01:36:04 - clean up volumes",
                                "2026-03-21 01:36:04 - disconnect routers",
                                "2026-03-21 01:36:04 - clean up subnets",
                                "2026-03-21 01:36:04 - clean up networks",
                                "2026-03-21 01:36:04 - clean up security groups",
                                "2026-03-21 01:36:04 - clean up floating ips",
                                "2026-03-21 01:36:04 - clean up routers"
                            ],
                            "zuul_log_id": "0242ac1f-6606-c9d8-0c48-000000000006-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:36:05.001563Z",
                            "start": "2026-03-21T01:36:02.488449Z"
                        },
                        "id": "0242ac1f-6606-c9d8-0c48-000000000006",
                        "name": "Clean the cloud environment"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 1,
            "failures": 0,
            "ignored": 0,
            "ok": 2,
            "rescued": 0,
            "skipped": 1,
            "unreachable": 0
        }
    },
    "trusted": false
},
{
    "branch": "main",
    "index": "2",
    "phase": "post",
    "playbook": "github.com/osism/zuul-config/playbooks/base/post-fetch.yaml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-03-21T01:36:09.692361Z",
                    "start": "2026-03-21T01:36:05.955249Z"
                },
                "id": "0242ac1f-6606-96e8-12bf-000000000002",
                "name": "Base post-fetch"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "changed": false,
                            "false_condition": "groups['all'] | length > 1",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-96e8-12bf-000000000006",
                        "name": "fetch-output",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/fetch-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:36:06.038637Z",
                            "start": "2026-03-21T01:36:05.972114Z"
                        },
                        "id": "0242ac1f-6606-96e8-12bf-000000000008",
                        "name": "Set log path for multiple nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "log_path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/logs"
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-96e8-12bf-000000000006",
                        "name": "fetch-output",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/fetch-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:36:06.104716Z",
                            "start": "2026-03-21T01:36:06.053339Z"
                        },
                        "id": "0242ac1f-6606-96e8-12bf-000000000009",
                        "name": "Set log path for single node"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": true,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_output_dir",
                                    "changed": false,
                                    "diff": {
                                        "after": {
                                            "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/logs"
                                        },
                                        "before": {
                                            "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/logs"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 0,
                                    "group": "root",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/logs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "root",
                                    "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/logs",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 0,
                                    "zj_output_dir": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/logs"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dir",
                                    "changed": true,
                                    "diff": {
                                        "after": {
                                            "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/artifacts",
                                            "state": "directory"
                                        },
                                        "before": {
                                            "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/artifacts",
                                            "state": "absent"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 0,
                                    "group": "root",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/artifacts",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "root",
                                    "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/artifacts",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 0,
                                    "zj_output_dir": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/artifacts"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dir",
                                    "changed": true,
                                    "diff": {
                                        "after": {
                                            "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/docs",
                                            "state": "directory"
                                        },
                                        "before": {
                                            "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/docs",
                                            "state": "absent"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 0,
                                    "group": "root",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/docs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "root",
                                    "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/docs",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 0,
                                    "zj_output_dir": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/docs"
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-96e8-12bf-000000000006",
                        "name": "fetch-output",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/fetch-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:36:07.198783Z",
                            "start": "2026-03-21T01:36:06.113422Z"
                        },
                        "id": "0242ac1f-6606-96e8-12bf-00000000000a",
                        "name": "Ensure local output dirs"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "synchronize",
                            "changed": true,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_output",
                                    "changed": true,
                                    "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --no-owner --no-group --rsh='/usr/bin/ssh -S none -o Port=22' --out-format='<<CHANGED>>%i %n%L' zuul-testbed04@81.163.193.140:/home/zuul-testbed04/zuul-output/logs/ /var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/logs/",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_local_rsync_password": null,
                                            "_local_rsync_path": "rsync",
                                            "_ssh_args": null,
                                            "_substitute_controller": false,
                                            "archive": true,
                                            "checksum": false,
                                            "compress": true,
                                            "copy_links": false,
                                            "delay_updates": true,
                                            "delete": false,
                                            "dest": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/logs/",
                                            "dest_port": 22,
                                            "dirs": false,
                                            "existing_only": false,
                                            "group": false,
                                            "link_dest": null,
                                            "links": null,
                                            "mode": "pull",
                                            "owner": false,
                                            "partial": false,
                                            "perms": null,
                                            "private_key": null,
                                            "recursive": null,
                                            "rsync_opts": [],
                                            "rsync_path": null,
                                            "rsync_timeout": 0,
                                            "set_remote_user": true,
                                            "src": "zuul-testbed04@81.163.193.140:/home/zuul-testbed04/zuul-output/logs/",
                                            "ssh_connection_multiplexing": false,
                                            "times": null,
                                            "use_ssh_args": false,
                                            "verify_host": true
                                        }
                                    },
                                    "msg": ".d..t...... ./\n",
                                    "rc": 0,
                                    "stdout_lines": [
                                        ".d..t...... ./"
                                    ],
                                    "zj_output": {
                                        "dest": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/logs",
                                        "src": "logs"
                                    }
                                },
                                {
                                    "ansible_loop_var": "zj_output",
                                    "changed": true,
                                    "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --no-owner --no-group --rsh='/usr/bin/ssh -S none -o Port=22' --out-format='<<CHANGED>>%i %n%L' zuul-testbed04@81.163.193.140:/home/zuul-testbed04/zuul-output/artifacts/ /var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/artifacts/",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_local_rsync_password": null,
                                            "_local_rsync_path": "rsync",
                                            "_ssh_args": null,
                                            "_substitute_controller": false,
                                            "archive": true,
                                            "checksum": false,
                                            "compress": true,
                                            "copy_links": false,
                                            "delay_updates": true,
                                            "delete": false,
                                            "dest": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/artifacts/",
                                            "dest_port": 22,
                                            "dirs": false,
                                            "existing_only": false,
                                            "group": false,
                                            "link_dest": null,
                                            "links": null,
                                            "mode": "pull",
                                            "owner": false,
                                            "partial": false,
                                            "perms": null,
                                            "private_key": null,
                                            "recursive": null,
                                            "rsync_opts": [],
                                            "rsync_path": null,
                                            "rsync_timeout": 0,
                                            "set_remote_user": true,
                                            "src": "zuul-testbed04@81.163.193.140:/home/zuul-testbed04/zuul-output/artifacts/",
                                            "ssh_connection_multiplexing": false,
                                            "times": null,
                                            "use_ssh_args": false,
                                            "verify_host": true
                                        }
                                    },
                                    "msg": ".d..t...... ./\n",
                                    "rc": 0,
                                    "stdout_lines": [
                                        ".d..t...... ./"
                                    ],
                                    "zj_output": {
                                        "dest": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/artifacts",
                                        "src": "artifacts"
                                    }
                                },
                                {
                                    "ansible_loop_var": "zj_output",
                                    "changed": true,
                                    "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --no-owner --no-group --rsh='/usr/bin/ssh -S none -o Port=22' --out-format='<<CHANGED>>%i %n%L' zuul-testbed04@81.163.193.140:/home/zuul-testbed04/zuul-output/docs/ /var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/docs/",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_local_rsync_password": null,
                                            "_local_rsync_path": "rsync",
                                            "_ssh_args": null,
                                            "_substitute_controller": false,
                                            "archive": true,
                                            "checksum": false,
                                            "compress": true,
                                            "copy_links": false,
                                            "delay_updates": true,
                                            "delete": false,
                                            "dest": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/docs/",
                                            "dest_port": 22,
                                            "dirs": false,
                                            "existing_only": false,
                                            "group": false,
                                            "link_dest": null,
                                            "links": null,
                                            "mode": "pull",
                                            "owner": false,
                                            "partial": false,
                                            "perms": null,
                                            "private_key": null,
                                            "recursive": null,
                                            "rsync_opts": [],
                                            "rsync_path": null,
                                            "rsync_timeout": 0,
                                            "set_remote_user": true,
                                            "src": "zuul-testbed04@81.163.193.140:/home/zuul-testbed04/zuul-output/docs/",
                                            "ssh_connection_multiplexing": false,
                                            "times": null,
                                            "use_ssh_args": false,
                                            "verify_host": true
                                        }
                                    },
                                    "msg": ".d..t...... ./\n",
                                    "rc": 0,
                                    "stdout_lines": [
                                        ".d..t...... ./"
                                    ],
                                    "zj_output": {
                                        "dest": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/work/docs",
                                        "src": "docs"
                                    }
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-96e8-12bf-000000000006",
                        "name": "fetch-output",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/fetch-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:36:09.622236Z",
                            "start": "2026-03-21T01:36:07.212765Z"
                        },
                        "id": "0242ac1f-6606-96e8-12bf-00000000000c",
                        "name": "Collect logs, artifacts and docs"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "shell",
                            "changed": false,
                            "msg": "All items skipped",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_item",
                                    "changed": false,
                                    "false_condition": "zuul.change is defined",
                                    "skip_reason": "Conditional result was False",
                                    "skipped": true,
                                    "zj_item": "artifacts"
                                },
                                {
                                    "ansible_loop_var": "zj_item",
                                    "changed": false,
                                    "false_condition": "zuul.change is defined",
                                    "skip_reason": "Conditional result was False",
                                    "skipped": true,
                                    "zj_item": "docs"
                                }
                            ],
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-96e8-12bf-00000000000f",
                        "name": "merge-output-to-logs",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/merge-output-to-logs"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:36:09.692361Z",
                            "start": "2026-03-21T01:36:09.646560Z"
                        },
                        "id": "0242ac1f-6606-96e8-12bf-000000000011",
                        "name": "Move artifacts and docs to logs dir"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 2,
            "failures": 0,
            "ignored": 0,
            "ok": 3,
            "rescued": 0,
            "skipped": 2,
            "unreachable": 0
        }
    },
    "trusted": true
},
{
    "branch": "main",
    "index": "3",
    "phase": "post",
    "playbook": "github.com/osism/zuul-config/playbooks/base/post.yaml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-03-21T01:36:11.651525Z",
                    "start": "2026-03-21T01:36:10.603257Z"
                },
                "id": "0242ac1f-6606-455b-2707-000000000002",
                "name": "Base post"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "authorized_key",
                            "changed": true,
                            "comment": null,
                            "exclusive": false,
                            "follow": false,
                            "invocation": {
                                "module_args": {
                                    "changed": true,
                                    "comment": null,
                                    "exclusive": false,
                                    "follow": false,
                                    "key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDiFU4X6bBG4GOayutbGDh+9mZMjWvhGFfP2Ynam+67TPObfnNalmbdJk06dNX/TYaC+yECnjjU2QtjNemgB1Of/27WgEMDtf2N+7DmrkWuvstUf0h1Qz2h5b0CxoqRCje/cCp4yqEFtqUuGRjZoGYEFZBIN1kvfPDNLFPyPYPfcxEgMtuwS87HTBdyWTHJu2NYPKyVjSLKpkiwtL19MyrYhjkD+Xlvaf6BJFCpdIbxa3muVVRkFr4vLWC+Rhu79Fk4XCy6MvyhU1VkAiv7qvjfE8KyeVnDvLy5nWYfHtyIa5rPDtwkoeHJdfsf4UN0qm2oNOHlMcD1fmg3jDUdDjpXYy8hLhiC0qKMRIiRTokNU+zw1rTDJSr78tAzap9ouJ6D0Tq10mqBZGUEpyM7n37IFOPU2fQmeoSvGQiWIl0dporJnK+QmiDEbs742C3DFQaEPXaJ5K4Gm6LcxeGgsqCzFXjGPOtIG98tST2dNvyzJX+xoi/i7/ALp03+FFZccaM= zuul-build-sshkey",
                                    "key_options": null,
                                    "keyfile": "/home/zuul-testbed04/.ssh/authorized_keys",
                                    "manage_dir": true,
                                    "path": null,
                                    "state": "absent",
                                    "user": "zuul-testbed04",
                                    "validate_certs": true
                                }
                            },
                            "key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDiFU4X6bBG4GOayutbGDh+9mZMjWvhGFfP2Ynam+67TPObfnNalmbdJk06dNX/TYaC+yECnjjU2QtjNemgB1Of/27WgEMDtf2N+7DmrkWuvstUf0h1Qz2h5b0CxoqRCje/cCp4yqEFtqUuGRjZoGYEFZBIN1kvfPDNLFPyPYPfcxEgMtuwS87HTBdyWTHJu2NYPKyVjSLKpkiwtL19MyrYhjkD+Xlvaf6BJFCpdIbxa3muVVRkFr4vLWC+Rhu79Fk4XCy6MvyhU1VkAiv7qvjfE8KyeVnDvLy5nWYfHtyIa5rPDtwkoeHJdfsf4UN0qm2oNOHlMcD1fmg3jDUdDjpXYy8hLhiC0qKMRIiRTokNU+zw1rTDJSr78tAzap9ouJ6D0Tq10mqBZGUEpyM7n37IFOPU2fQmeoSvGQiWIl0dporJnK+QmiDEbs742C3DFQaEPXaJ5K4Gm6LcxeGgsqCzFXjGPOtIG98tST2dNvyzJX+xoi/i7/ALp03+FFZccaM= zuul-build-sshkey",
                            "key_options": null,
                            "keyfile": "/home/zuul-testbed04/.ssh/authorized_keys",
                            "manage_dir": true,
                            "path": null,
                            "state": "absent",
                            "user": "zuul-testbed04",
                            "validate_certs": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-455b-2707-000000000005",
                        "name": "remove-build-sshkey",
                        "path": "/var/lib/zuul/builds/a3240bd8b4c74e6f8c9446925cc2fed6/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/remove-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-21T01:36:11.651525Z",
                            "start": "2026-03-21T01:36:10.618305Z"
                        },
                        "id": "0242ac1f-6606-455b-2707-000000000007",
                        "name": "Remove the build SSH key from all nodes"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 1,
            "failures": 0,
            "ignored": 0,
            "ok": 1,
            "rescued": 0,
            "skipped": 0,
            "unreachable": 0
        }
    },
    "trusted": true
}
]
